chore(exporter): rework exporter to fix wrong histogram usage and cache metric data
This commit is contained in:
parent
4d2a7acebc
commit
8b7f45563a
6 changed files with 180 additions and 312 deletions
|
@ -17,6 +17,7 @@ type AppSettings struct {
|
|||
ReadHeaderTimeout time.Duration
|
||||
EnableExporter bool
|
||||
ExporterAddress string
|
||||
ExporterInterval time.Duration
|
||||
}
|
||||
|
||||
//nolint:lll // ignore line length
|
||||
|
@ -28,6 +29,7 @@ type CLI struct {
|
|||
ReadHeaderTimeout string `name:"read-header-timeout" env:"GEOIP_READ_HEADER_TIMEOUT" help:"timeout for reading http header" default:"${default_read_header_timeout}"`
|
||||
EnableExporter bool `name:"enable-exporter" env:"GEOIP_ENABLE_EXPORTER" help:"enable prometheus exporter" default:"${default_enable_exporter}"`
|
||||
ExporterAddress string `name:"exporter-address" env:"GEOIP_EXPORTER_ADDRESS" help:"Address to use for the prometheus metrics server" default:"${default_exporter_address}"`
|
||||
ExporterInterval string `name:"exporter-interval" env:"GEOIP_EXPORTER_INTERVAL" help:"Interval to scrape the new metrics data" default:"${default_exporter_interval}"`
|
||||
}
|
||||
|
||||
func (c *CLI) Parse() (*AppSettings, error) {
|
||||
|
@ -40,6 +42,7 @@ func (c *CLI) Parse() (*AppSettings, error) {
|
|||
"default_read_header_timeout": "3s",
|
||||
"default_enable_exporter": "false",
|
||||
"default_exporter_address": ":9191",
|
||||
"default_exporter_interval": "5s",
|
||||
},
|
||||
kong.Name("country-geo-locations"),
|
||||
kong.Description("🚀 Start a simple web server for GeoIP data"),
|
||||
|
@ -56,6 +59,11 @@ func (c *CLI) Parse() (*AppSettings, error) {
|
|||
return nil, err
|
||||
}
|
||||
|
||||
exporterInterval, err := time.ParseDuration(c.ExporterInterval)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
if c.DataURL == "" {
|
||||
return nil, ErrInvalidDataURL
|
||||
}
|
||||
|
@ -68,5 +76,6 @@ func (c *CLI) Parse() (*AppSettings, error) {
|
|||
ReadHeaderTimeout: readHeaderTimeout,
|
||||
EnableExporter: c.EnableExporter,
|
||||
ExporterAddress: c.ExporterAddress,
|
||||
ExporterInterval: exporterInterval,
|
||||
}, nil
|
||||
}
|
||||
|
|
|
@ -1,46 +0,0 @@
|
|||
package exporter
|
||||
|
||||
import (
|
||||
"git.ar21.de/yolokube/country-geo-locations/internal/cache"
|
||||
"git.ar21.de/yolokube/country-geo-locations/internal/cmd"
|
||||
"git.ar21.de/yolokube/country-geo-locations/internal/database"
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
)
|
||||
|
||||
type Collector struct {
|
||||
config *cmd.AppSettings
|
||||
cache *cache.Cache
|
||||
db *database.Database
|
||||
metrics *Metrics
|
||||
queue *RequestDataQueue
|
||||
}
|
||||
|
||||
func NewCollector(
|
||||
config *cmd.AppSettings,
|
||||
cache *cache.Cache,
|
||||
db *database.Database,
|
||||
queue *RequestDataQueue,
|
||||
) *Collector {
|
||||
return &Collector{
|
||||
config: config,
|
||||
cache: cache,
|
||||
db: db,
|
||||
metrics: NewMetrics(),
|
||||
queue: queue,
|
||||
}
|
||||
}
|
||||
|
||||
func (c *Collector) Describe(ch chan<- *prometheus.Desc) {
|
||||
ch <- c.metrics.metricCacheTTL
|
||||
ch <- c.metrics.metricCurrentlyCached
|
||||
ch <- c.metrics.metricDatabaseTimestamp
|
||||
ch <- c.metrics.metricDatabaseReady
|
||||
}
|
||||
|
||||
func (c *Collector) Collect(ch chan<- prometheus.Metric) {
|
||||
c.metrics.collectCacheTTLMetric(ch, c.config.CacheTTL.Seconds())
|
||||
c.metrics.collectCurrentlyCachedMetric(ch, float64(c.cache.Count()))
|
||||
c.metrics.collectDatabaseTimestampMetric(ch, c.db)
|
||||
c.metrics.collectDatabaseReadyMetric(ch, c.db.IsReady())
|
||||
c.metrics.collectReqeustDataMetrics(ch, c.queue)
|
||||
}
|
61
internal/exporter/exporter.go
Normal file
61
internal/exporter/exporter.go
Normal file
|
@ -0,0 +1,61 @@
|
|||
package exporter
|
||||
|
||||
import (
|
||||
"net/http"
|
||||
|
||||
"git.ar21.de/yolokube/country-geo-locations/internal/cache"
|
||||
"git.ar21.de/yolokube/country-geo-locations/internal/cmd"
|
||||
"git.ar21.de/yolokube/country-geo-locations/internal/database"
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
"github.com/prometheus/client_golang/prometheus/promhttp"
|
||||
)
|
||||
|
||||
type Exporter struct {
|
||||
config *cmd.AppSettings
|
||||
cache *cache.Cache
|
||||
database *database.Database
|
||||
metrics *Metrics
|
||||
}
|
||||
|
||||
func NewExporter(config *cmd.AppSettings, c *cache.Cache, db *database.Database) *Exporter {
|
||||
exporter := &Exporter{
|
||||
config: config,
|
||||
cache: c,
|
||||
database: db,
|
||||
}
|
||||
|
||||
exporter.metrics = NewMetrics(exporter)
|
||||
return exporter
|
||||
}
|
||||
|
||||
func (e *Exporter) Collect() {
|
||||
e.metrics.collectCacheTTLMetric()
|
||||
e.metrics.collectCurrentlyCachedMetric()
|
||||
e.metrics.collectDatabaseReadyMetric()
|
||||
e.metrics.collectDatabaseTimestampMetric()
|
||||
}
|
||||
|
||||
func (e *Exporter) Middleware() func(next http.Handler) http.Handler {
|
||||
return Middleware{
|
||||
metrics: e.metrics,
|
||||
}.handler
|
||||
}
|
||||
|
||||
func (e *Exporter) Start() error {
|
||||
prometheus.MustRegister(
|
||||
e.metrics.metricCacheTTL,
|
||||
e.metrics.metricCurrentlyCached,
|
||||
e.metrics.metricDatabaseTimestamp,
|
||||
e.metrics.metricDatabaseReady,
|
||||
e.metrics.metricRequestsTotal,
|
||||
e.metrics.metricRequestLatency,
|
||||
)
|
||||
|
||||
server := &http.Server{
|
||||
Addr: e.config.ExporterAddress,
|
||||
Handler: promhttp.Handler(),
|
||||
ReadHeaderTimeout: e.config.ReadHeaderTimeout,
|
||||
}
|
||||
|
||||
return server.ListenAndServe()
|
||||
}
|
|
@ -2,9 +2,7 @@ package exporter
|
|||
|
||||
import (
|
||||
"log"
|
||||
"time"
|
||||
|
||||
"git.ar21.de/yolokube/country-geo-locations/internal/database"
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
)
|
||||
|
||||
|
@ -13,205 +11,102 @@ const (
|
|||
cacheSubsystem string = "cache"
|
||||
dbSubsystem string = "db"
|
||||
|
||||
metricLabelCacheTTL string = "ttl"
|
||||
metricLabelCurrentlyCached string = "currently_cached"
|
||||
metricLabelDatabaseTimestamp string = "timestamp"
|
||||
metricLabelDatabaseReady string = "ready"
|
||||
metricLabelRequestsTotal string = "requests_total"
|
||||
metricLabelRequestLatency string = "request_latency"
|
||||
metricNameCacheTTL string = "ttl"
|
||||
metricNameCurrentlyCached string = "currently_cached"
|
||||
metricNameDatabaseTimestamp string = "timestamp"
|
||||
metricNameDatabaseReady string = "ready"
|
||||
metricNameRequestsTotal string = "requests_total"
|
||||
metricNameRequestLatency string = "request_latency"
|
||||
)
|
||||
|
||||
type Metrics struct {
|
||||
metricCacheTTL *prometheus.Desc
|
||||
metricCurrentlyCached *prometheus.Desc
|
||||
metricDatabaseTimestamp *prometheus.Desc
|
||||
metricDatabaseReady *prometheus.Desc
|
||||
metricRequestsTotal *prometheus.Desc
|
||||
metricRequestLatency *prometheus.Desc
|
||||
exporter *Exporter
|
||||
|
||||
counter uint
|
||||
metricCacheTTL prometheus.Gauge
|
||||
metricCurrentlyCached prometheus.Gauge
|
||||
metricDatabaseTimestamp *prometheus.GaugeVec
|
||||
metricDatabaseReady prometheus.Gauge
|
||||
metricRequestsTotal prometheus.Counter
|
||||
metricRequestLatency prometheus.Histogram
|
||||
}
|
||||
|
||||
func NewMetrics() *Metrics {
|
||||
func NewMetrics(e *Exporter) *Metrics {
|
||||
return &Metrics{
|
||||
metricCacheTTL: prometheus.NewDesc(
|
||||
prometheus.BuildFQName(
|
||||
namespace,
|
||||
cacheSubsystem,
|
||||
metricLabelCacheTTL,
|
||||
),
|
||||
"Duration for cached requests",
|
||||
nil,
|
||||
nil,
|
||||
exporter: e,
|
||||
metricCacheTTL: prometheus.NewGauge(
|
||||
prometheus.GaugeOpts{
|
||||
Namespace: namespace,
|
||||
Subsystem: cacheSubsystem,
|
||||
Name: metricNameCacheTTL,
|
||||
Help: "Duration for cached requests",
|
||||
},
|
||||
),
|
||||
metricCurrentlyCached: prometheus.NewDesc(
|
||||
prometheus.BuildFQName(
|
||||
namespace,
|
||||
cacheSubsystem,
|
||||
metricLabelCurrentlyCached,
|
||||
),
|
||||
"Number of cached entries",
|
||||
nil,
|
||||
nil,
|
||||
metricCurrentlyCached: prometheus.NewGauge(
|
||||
prometheus.GaugeOpts{
|
||||
Namespace: namespace,
|
||||
Subsystem: cacheSubsystem,
|
||||
Name: metricNameCurrentlyCached,
|
||||
Help: "Number of cached entries",
|
||||
},
|
||||
),
|
||||
metricDatabaseTimestamp: prometheus.NewDesc(
|
||||
prometheus.BuildFQName(
|
||||
namespace,
|
||||
dbSubsystem,
|
||||
metricLabelDatabaseTimestamp,
|
||||
),
|
||||
"Timestamp of the CSV file",
|
||||
[]string{metricLabelDatabaseTimestamp},
|
||||
nil,
|
||||
metricDatabaseTimestamp: prometheus.NewGaugeVec(
|
||||
prometheus.GaugeOpts{
|
||||
Namespace: namespace,
|
||||
Subsystem: dbSubsystem,
|
||||
Name: metricNameDatabaseTimestamp,
|
||||
Help: "Timestamp of the CSV file",
|
||||
},
|
||||
[]string{metricNameDatabaseTimestamp},
|
||||
),
|
||||
metricDatabaseReady: prometheus.NewDesc(
|
||||
prometheus.BuildFQName(
|
||||
namespace,
|
||||
dbSubsystem,
|
||||
metricLabelDatabaseReady,
|
||||
),
|
||||
"Ready status of the database",
|
||||
nil,
|
||||
nil,
|
||||
metricDatabaseReady: prometheus.NewGauge(
|
||||
prometheus.GaugeOpts{
|
||||
Namespace: namespace,
|
||||
Subsystem: dbSubsystem,
|
||||
Name: metricNameDatabaseReady,
|
||||
Help: "Ready status of the database",
|
||||
},
|
||||
),
|
||||
metricRequestsTotal: prometheus.NewDesc(
|
||||
prometheus.BuildFQName(
|
||||
namespace,
|
||||
"",
|
||||
metricLabelRequestsTotal,
|
||||
),
|
||||
"Counter for total requests",
|
||||
nil,
|
||||
nil,
|
||||
metricRequestsTotal: prometheus.NewCounter(
|
||||
prometheus.CounterOpts{
|
||||
Namespace: namespace,
|
||||
Name: metricNameRequestsTotal,
|
||||
Help: "Counter for total requests",
|
||||
},
|
||||
),
|
||||
metricRequestLatency: prometheus.NewDesc(
|
||||
prometheus.BuildFQName(
|
||||
namespace,
|
||||
"",
|
||||
metricLabelRequestLatency,
|
||||
),
|
||||
"Latency statistics for requests",
|
||||
nil,
|
||||
nil,
|
||||
metricRequestLatency: prometheus.NewHistogram(
|
||||
prometheus.HistogramOpts{
|
||||
Namespace: namespace,
|
||||
Name: metricNameRequestLatency,
|
||||
Help: "Latency statistics for requests",
|
||||
Buckets: prometheus.ExponentialBuckets(10, 1.5, 30),
|
||||
},
|
||||
),
|
||||
}
|
||||
}
|
||||
|
||||
func (m *Metrics) collectCacheTTLMetric(ch chan<- prometheus.Metric, ttl float64) {
|
||||
ch <- prometheus.MustNewConstMetric(
|
||||
m.metricCacheTTL,
|
||||
prometheus.GaugeValue,
|
||||
ttl,
|
||||
)
|
||||
func (m *Metrics) collectCacheTTLMetric() {
|
||||
m.metricCacheTTL.Set(m.exporter.config.CacheTTL.Seconds())
|
||||
}
|
||||
|
||||
func (m *Metrics) collectCurrentlyCachedMetric(ch chan<- prometheus.Metric, count float64) {
|
||||
ch <- prometheus.MustNewConstMetric(
|
||||
m.metricCurrentlyCached,
|
||||
prometheus.GaugeValue,
|
||||
count,
|
||||
)
|
||||
func (m *Metrics) collectCurrentlyCachedMetric() {
|
||||
m.metricCurrentlyCached.Set(float64(m.exporter.cache.Count()))
|
||||
}
|
||||
|
||||
func (m *Metrics) collectDatabaseTimestampMetric(ch chan<- prometheus.Metric, db *database.Database) {
|
||||
timestamp, err := db.Timestamp()
|
||||
if err == nil {
|
||||
ch <- prometheus.MustNewConstMetric(
|
||||
m.metricDatabaseTimestamp,
|
||||
prometheus.GaugeValue,
|
||||
float64(timestamp.Unix()),
|
||||
timestamp.String(),
|
||||
)
|
||||
} else {
|
||||
log.Printf("failed to read file timestamp: %v\n", err)
|
||||
func (m *Metrics) collectDatabaseTimestampMetric() {
|
||||
timestamp, err := m.exporter.database.Timestamp()
|
||||
if err != nil {
|
||||
log.Printf("failed to read file timestamp: %v", err)
|
||||
return
|
||||
}
|
||||
|
||||
m.metricDatabaseTimestamp.WithLabelValues(timestamp.String()).Set(float64(timestamp.Unix()))
|
||||
}
|
||||
|
||||
func (m *Metrics) collectDatabaseReadyMetric(ch chan<- prometheus.Metric, ready bool) {
|
||||
var dbReady uint8
|
||||
if ready {
|
||||
func (m *Metrics) collectDatabaseReadyMetric() {
|
||||
var dbReady float64
|
||||
if m.exporter.database.IsReady() {
|
||||
dbReady = 1
|
||||
}
|
||||
|
||||
ch <- prometheus.MustNewConstMetric(
|
||||
m.metricDatabaseReady,
|
||||
prometheus.GaugeValue,
|
||||
float64(dbReady),
|
||||
)
|
||||
}
|
||||
|
||||
func (m *Metrics) collectReqeustDataMetrics(ch chan<- prometheus.Metric, queue *RequestDataQueue) {
|
||||
var (
|
||||
count uint64
|
||||
sum float64
|
||||
)
|
||||
buckets := make(map[float64]uint64)
|
||||
bucketBounds := []float64{
|
||||
10,
|
||||
20,
|
||||
30,
|
||||
40,
|
||||
50,
|
||||
60,
|
||||
70,
|
||||
80,
|
||||
90,
|
||||
100,
|
||||
200,
|
||||
300,
|
||||
400,
|
||||
500,
|
||||
600,
|
||||
700,
|
||||
800,
|
||||
900,
|
||||
1000,
|
||||
1500,
|
||||
2000,
|
||||
2500,
|
||||
3000,
|
||||
3500,
|
||||
4000,
|
||||
4500,
|
||||
5000,
|
||||
10000,
|
||||
20000,
|
||||
30000,
|
||||
40000,
|
||||
50000,
|
||||
100000,
|
||||
200000,
|
||||
300000,
|
||||
400000,
|
||||
500000,
|
||||
1000000,
|
||||
}
|
||||
|
||||
data := queue.ConsumeAll()
|
||||
for _, r := range data {
|
||||
latency := float64(r.Latency.Microseconds())
|
||||
sum += latency
|
||||
count++
|
||||
|
||||
for _, bound := range bucketBounds {
|
||||
if latency <= bound {
|
||||
buckets[bound]++
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
m.counter += uint(len(data))
|
||||
ch <- prometheus.MustNewConstMetric(
|
||||
m.metricRequestsTotal,
|
||||
prometheus.CounterValue,
|
||||
float64(m.counter),
|
||||
)
|
||||
|
||||
ch <- prometheus.MustNewConstHistogramWithCreatedTimestamp(
|
||||
m.metricRequestLatency,
|
||||
count,
|
||||
sum,
|
||||
buckets,
|
||||
time.Now(),
|
||||
)
|
||||
m.metricDatabaseReady.Set(dbReady)
|
||||
}
|
||||
|
|
|
@ -2,67 +2,19 @@ package exporter
|
|||
|
||||
import (
|
||||
"net/http"
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
"github.com/go-chi/chi/v5/middleware"
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
)
|
||||
|
||||
type RequestData struct {
|
||||
Latency time.Duration
|
||||
Request *http.Request
|
||||
Start time.Time
|
||||
}
|
||||
|
||||
type RequestDataQueue struct {
|
||||
mu sync.Mutex
|
||||
data []RequestData
|
||||
}
|
||||
|
||||
func NewRequestDataQueue() *RequestDataQueue {
|
||||
return &RequestDataQueue{
|
||||
data: []RequestData{},
|
||||
}
|
||||
}
|
||||
|
||||
func (q *RequestDataQueue) Add(data RequestData) {
|
||||
q.mu.Lock()
|
||||
defer q.mu.Unlock()
|
||||
q.data = append(q.data, data)
|
||||
}
|
||||
|
||||
func (q *RequestDataQueue) ConsumeAll() []RequestData {
|
||||
q.mu.Lock()
|
||||
defer q.mu.Unlock()
|
||||
data := q.data
|
||||
q.data = nil
|
||||
return data
|
||||
}
|
||||
|
||||
type Middleware struct {
|
||||
queue *RequestDataQueue
|
||||
}
|
||||
|
||||
func NewMiddleware(queue *RequestDataQueue) func(next http.Handler) http.Handler {
|
||||
m := Middleware{
|
||||
queue: queue,
|
||||
}
|
||||
return m.handler
|
||||
metrics *Metrics
|
||||
}
|
||||
|
||||
func (m Middleware) handler(next http.Handler) http.Handler {
|
||||
fn := func(w http.ResponseWriter, r *http.Request) {
|
||||
start := time.Now()
|
||||
ww := middleware.NewWrapResponseWriter(w, r.ProtoMajor)
|
||||
next.ServeHTTP(ww, r)
|
||||
|
||||
m.queue.Add(
|
||||
RequestData{
|
||||
Latency: time.Since(start),
|
||||
Request: r,
|
||||
Start: start,
|
||||
},
|
||||
)
|
||||
}
|
||||
return http.HandlerFunc(fn)
|
||||
return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||
timer := prometheus.NewTimer(m.metrics.metricRequestLatency)
|
||||
next.ServeHTTP(w, r)
|
||||
m.metrics.metricRequestsTotal.Inc()
|
||||
timer.ObserveDuration()
|
||||
})
|
||||
}
|
||||
|
|
63
main.go
63
main.go
|
@ -6,6 +6,7 @@ import (
|
|||
"os"
|
||||
"os/signal"
|
||||
"syscall"
|
||||
"time"
|
||||
|
||||
apiv1 "git.ar21.de/yolokube/country-geo-locations/api/v1"
|
||||
"git.ar21.de/yolokube/country-geo-locations/internal/cache"
|
||||
|
@ -17,29 +18,24 @@ import (
|
|||
"github.com/go-chi/chi/v5"
|
||||
"github.com/go-chi/chi/v5/middleware"
|
||||
"github.com/go-chi/render"
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
"github.com/prometheus/client_golang/prometheus/promhttp"
|
||||
)
|
||||
|
||||
func main() {
|
||||
cli := cmd.CLI{}
|
||||
queue := exporter.NewRequestDataQueue()
|
||||
appSettings, err := cli.Parse()
|
||||
config, err := cli.Parse()
|
||||
if err != nil {
|
||||
log.Fatal(err)
|
||||
}
|
||||
|
||||
handleGracefulShutdown()
|
||||
|
||||
exporterMiddleware := exporter.NewMiddleware(queue)
|
||||
r := chi.NewRouter()
|
||||
r.Use(middleware.RequestID)
|
||||
r.Use(middleware.Logger)
|
||||
r.Use(middleware.Recoverer)
|
||||
r.Use(exporterMiddleware)
|
||||
r.Use(render.SetContentType(render.ContentTypeJSON))
|
||||
|
||||
ctx := downloader.NewContext(appSettings.DataFile, appSettings.DataURL)
|
||||
ctx := downloader.NewContext(config.DataFile, config.DataURL)
|
||||
if !ctx.FileExists() {
|
||||
if downloadErr := ctx.Download(); downloadErr != nil {
|
||||
log.Fatal(downloadErr)
|
||||
|
@ -47,24 +43,42 @@ func main() {
|
|||
log.Printf("saved file to %s\n", ctx.Filename)
|
||||
}
|
||||
|
||||
cache := cache.NewCache(appSettings.CacheTTL)
|
||||
db, err := database.NewDatabase(appSettings)
|
||||
cache := cache.NewCache(config.CacheTTL)
|
||||
db, err := database.NewDatabase(config)
|
||||
if err != nil {
|
||||
log.Fatal("database creation failed", err)
|
||||
}
|
||||
|
||||
if appSettings.EnableExporter {
|
||||
if config.EnableExporter {
|
||||
exporter := exporter.NewExporter(config, cache, db)
|
||||
r.Use(exporter.Middleware())
|
||||
|
||||
ticker := time.NewTicker(config.ExporterInterval)
|
||||
exit := make(chan struct{})
|
||||
go func() {
|
||||
err = enableExporter(appSettings, cache, db, queue)
|
||||
for {
|
||||
select {
|
||||
case <-ticker.C:
|
||||
exporter.Collect()
|
||||
case <-exit:
|
||||
ticker.Stop()
|
||||
return
|
||||
}
|
||||
}
|
||||
}()
|
||||
log.Println("prometheus exporter refreshes metric data every", config.ExporterInterval)
|
||||
|
||||
go func() {
|
||||
err = exporter.Start()
|
||||
if err != nil {
|
||||
log.Panic(err)
|
||||
}
|
||||
}()
|
||||
log.Println("prometheus exporter started at", appSettings.ExporterAddress)
|
||||
log.Println("prometheus exporter started at", config.ExporterAddress)
|
||||
}
|
||||
|
||||
log.Println("importing data from file", appSettings.DataFile)
|
||||
err = csvimporter.ImportCSV(appSettings.DataFile, db)
|
||||
log.Println("importing data from file", config.DataFile)
|
||||
err = csvimporter.ImportCSV(config.DataFile, db)
|
||||
if err != nil {
|
||||
log.Fatal("data Import from file failed", err)
|
||||
}
|
||||
|
@ -74,9 +88,9 @@ func main() {
|
|||
r.Mount("/api/v1", apiv1.NewRouter(lh))
|
||||
|
||||
server := &http.Server{
|
||||
Addr: appSettings.ServerAddress,
|
||||
Addr: config.ServerAddress,
|
||||
Handler: r,
|
||||
ReadHeaderTimeout: appSettings.ReadHeaderTimeout,
|
||||
ReadHeaderTimeout: config.ReadHeaderTimeout,
|
||||
}
|
||||
|
||||
log.Println("starting server at", server.Addr)
|
||||
|
@ -85,23 +99,6 @@ func main() {
|
|||
}
|
||||
}
|
||||
|
||||
func enableExporter(
|
||||
settings *cmd.AppSettings,
|
||||
cache *cache.Cache,
|
||||
db *database.Database,
|
||||
queue *exporter.RequestDataQueue,
|
||||
) error {
|
||||
prometheus.MustRegister(exporter.NewCollector(settings, cache, db, queue))
|
||||
|
||||
metricsServer := &http.Server{
|
||||
Addr: settings.ExporterAddress,
|
||||
Handler: promhttp.Handler(),
|
||||
ReadHeaderTimeout: settings.ReadHeaderTimeout,
|
||||
}
|
||||
|
||||
return metricsServer.ListenAndServe()
|
||||
}
|
||||
|
||||
func handleGracefulShutdown() {
|
||||
var signals = make(chan os.Signal, 1)
|
||||
|
||||
|
|
Loading…
Reference in a new issue