wrapper/monitoring/prometheus: fix multiservice wrapper usage
Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org>
This commit is contained in:
parent
9cb28538ec
commit
f2c41beb07
332
prometheus.go
332
prometheus.go
@ -3,6 +3,7 @@ package prometheus
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"sync"
|
||||||
|
|
||||||
"github.com/micro/go-micro/v2/client"
|
"github.com/micro/go-micro/v2/client"
|
||||||
"github.com/micro/go-micro/v2/logger"
|
"github.com/micro/go-micro/v2/logger"
|
||||||
@ -13,84 +14,121 @@ import (
|
|||||||
|
|
||||||
var (
|
var (
|
||||||
// default metric prefix
|
// default metric prefix
|
||||||
DefaultMetricPrefix = "micro"
|
DefaultMetricPrefix = "micro_"
|
||||||
// default label prefix
|
// default label prefix
|
||||||
DefaultLabelPrefix = "micro"
|
DefaultLabelPrefix = "micro_"
|
||||||
)
|
|
||||||
|
|
||||||
type wrapper struct {
|
|
||||||
opsCounter *prometheus.CounterVec
|
opsCounter *prometheus.CounterVec
|
||||||
timeCounterSummary *prometheus.SummaryVec
|
timeCounterSummary *prometheus.SummaryVec
|
||||||
timeCounterHistogram *prometheus.HistogramVec
|
timeCounterHistogram *prometheus.HistogramVec
|
||||||
callFunc client.CallFunc
|
|
||||||
client.Client
|
mu sync.Mutex
|
||||||
|
)
|
||||||
|
|
||||||
|
type Options struct {
|
||||||
|
Name string
|
||||||
|
Version string
|
||||||
|
ID string
|
||||||
}
|
}
|
||||||
|
|
||||||
func getLabels(opts ...server.Option) prometheus.Labels {
|
type Option func(*Options)
|
||||||
sopts := server.Options{}
|
|
||||||
|
|
||||||
for _, opt := range opts {
|
func ServiceName(name string) Option {
|
||||||
opt(&sopts)
|
return func(opts *Options) {
|
||||||
|
opts.Name = name
|
||||||
}
|
}
|
||||||
|
|
||||||
labels := make(prometheus.Labels, len(sopts.Metadata))
|
|
||||||
for k, v := range sopts.Metadata {
|
|
||||||
labels[fmt.Sprintf("%s_%s", DefaultLabelPrefix, k)] = v
|
|
||||||
}
|
|
||||||
if len(sopts.Name) > 0 {
|
|
||||||
labels[fmt.Sprintf("%s_%s", DefaultLabelPrefix, "name")] = sopts.Name
|
|
||||||
}
|
|
||||||
if len(sopts.Id) > 0 {
|
|
||||||
labels[fmt.Sprintf("%s_%s", DefaultLabelPrefix, "id")] = sopts.Id
|
|
||||||
}
|
|
||||||
if len(sopts.Version) > 0 {
|
|
||||||
labels[fmt.Sprintf("%s_%s", DefaultLabelPrefix, "version")] = sopts.Version
|
|
||||||
}
|
|
||||||
|
|
||||||
return labels
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewClientWrapper() client.Wrapper {
|
func ServiceVersion(version string) Option {
|
||||||
return func(c client.Client) client.Client {
|
return func(opts *Options) {
|
||||||
opsCounter := prometheus.NewCounterVec(
|
opts.Version = version
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func ServiceID(id string) Option {
|
||||||
|
return func(opts *Options) {
|
||||||
|
opts.ID = id
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func registerMetrics() {
|
||||||
|
mu.Lock()
|
||||||
|
defer mu.Unlock()
|
||||||
|
|
||||||
|
if opsCounter == nil {
|
||||||
|
opsCounter = prometheus.NewCounterVec(
|
||||||
prometheus.CounterOpts{
|
prometheus.CounterOpts{
|
||||||
Name: fmt.Sprintf("%s_client_request_total", DefaultMetricPrefix),
|
Name: fmt.Sprintf("%srequest_total", DefaultMetricPrefix),
|
||||||
Help: "How many requests called, partitioned by method and status",
|
Help: "Requests processed, partitioned by endpoint and status",
|
||||||
|
},
|
||||||
|
[]string{
|
||||||
|
fmt.Sprintf("%s%s", DefaultLabelPrefix, "name"),
|
||||||
|
fmt.Sprintf("%s%s", DefaultLabelPrefix, "version"),
|
||||||
|
fmt.Sprintf("%s%s", DefaultLabelPrefix, "id"),
|
||||||
|
fmt.Sprintf("%s%s", DefaultLabelPrefix, "endpoint"),
|
||||||
|
fmt.Sprintf("%s%s", DefaultLabelPrefix, "status"),
|
||||||
},
|
},
|
||||||
[]string{"method", "status"},
|
|
||||||
)
|
)
|
||||||
|
}
|
||||||
|
|
||||||
timeCounterSummary := prometheus.NewSummaryVec(
|
if timeCounterSummary == nil {
|
||||||
|
timeCounterSummary = prometheus.NewSummaryVec(
|
||||||
prometheus.SummaryOpts{
|
prometheus.SummaryOpts{
|
||||||
Name: fmt.Sprintf("%s_client_latency_microseconds", DefaultMetricPrefix),
|
Name: fmt.Sprintf("%slatency_microseconds", DefaultMetricPrefix),
|
||||||
Help: "Service client request latencies in microseconds",
|
Help: "Request latencies in microseconds, partitioned by endpoint",
|
||||||
|
},
|
||||||
|
[]string{
|
||||||
|
fmt.Sprintf("%s%s", DefaultLabelPrefix, "name"),
|
||||||
|
fmt.Sprintf("%s%s", DefaultLabelPrefix, "version"),
|
||||||
|
fmt.Sprintf("%s%s", DefaultLabelPrefix, "id"),
|
||||||
|
fmt.Sprintf("%s%s", DefaultLabelPrefix, "endpoint"),
|
||||||
},
|
},
|
||||||
[]string{"method"},
|
|
||||||
)
|
)
|
||||||
|
}
|
||||||
|
|
||||||
timeCounterHistogram := prometheus.NewHistogramVec(
|
if timeCounterHistogram == nil {
|
||||||
|
timeCounterHistogram = prometheus.NewHistogramVec(
|
||||||
prometheus.HistogramOpts{
|
prometheus.HistogramOpts{
|
||||||
Name: fmt.Sprintf("%s_client_request_duration_seconds", DefaultMetricPrefix),
|
Name: fmt.Sprintf("%srequest_duration_seconds", DefaultMetricPrefix),
|
||||||
Help: "Service client request time in seconds",
|
Help: "Request time in seconds, partitioned by endpoint",
|
||||||
|
},
|
||||||
|
[]string{
|
||||||
|
fmt.Sprintf("%s%s", DefaultLabelPrefix, "name"),
|
||||||
|
fmt.Sprintf("%s%s", DefaultLabelPrefix, "version"),
|
||||||
|
fmt.Sprintf("%s%s", DefaultLabelPrefix, "id"),
|
||||||
|
fmt.Sprintf("%s%s", DefaultLabelPrefix, "endpoint"),
|
||||||
},
|
},
|
||||||
[]string{"method"},
|
|
||||||
)
|
)
|
||||||
|
}
|
||||||
|
|
||||||
for _, collector := range []prometheus.Collector{opsCounter, timeCounterSummary, timeCounterHistogram} {
|
for _, collector := range []prometheus.Collector{opsCounter, timeCounterSummary, timeCounterHistogram} {
|
||||||
if err := prometheus.DefaultRegisterer.Register(collector); err != nil {
|
if err := prometheus.DefaultRegisterer.Register(collector); err != nil {
|
||||||
// if already registered, skip fatal
|
// if already registered, skip fatal
|
||||||
if are, ok := err.(prometheus.AlreadyRegisteredError); ok {
|
if _, ok := err.(prometheus.AlreadyRegisteredError); !ok {
|
||||||
collector = are.ExistingCollector
|
|
||||||
} else {
|
|
||||||
logger.Fatal(err)
|
logger.Fatal(err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
type wrapper struct {
|
||||||
|
options Options
|
||||||
|
callFunc client.CallFunc
|
||||||
|
client.Client
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewClientWrapper(opts ...Option) client.Wrapper {
|
||||||
|
registerMetrics()
|
||||||
|
|
||||||
|
options := Options{}
|
||||||
|
for _, opt := range opts {
|
||||||
|
opt(&options)
|
||||||
|
}
|
||||||
|
|
||||||
|
return func(c client.Client) client.Client {
|
||||||
handler := &wrapper{
|
handler := &wrapper{
|
||||||
timeCounterHistogram: timeCounterHistogram,
|
options: options,
|
||||||
timeCounterSummary: timeCounterSummary,
|
|
||||||
opsCounter: opsCounter,
|
|
||||||
Client: c,
|
Client: c,
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -98,47 +136,17 @@ func NewClientWrapper() client.Wrapper {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewCallWrapper() client.CallWrapper {
|
func NewCallWrapper(opts ...Option) client.CallWrapper {
|
||||||
|
registerMetrics()
|
||||||
|
|
||||||
|
options := Options{}
|
||||||
|
for _, opt := range opts {
|
||||||
|
opt(&options)
|
||||||
|
}
|
||||||
|
|
||||||
return func(fn client.CallFunc) client.CallFunc {
|
return func(fn client.CallFunc) client.CallFunc {
|
||||||
opsCounter := prometheus.NewCounterVec(
|
|
||||||
prometheus.CounterOpts{
|
|
||||||
Name: fmt.Sprintf("%s_call_request_total", DefaultMetricPrefix),
|
|
||||||
Help: "How many requests called, partitioned by method and status",
|
|
||||||
},
|
|
||||||
[]string{"method", "status"},
|
|
||||||
)
|
|
||||||
|
|
||||||
timeCounterSummary := prometheus.NewSummaryVec(
|
|
||||||
prometheus.SummaryOpts{
|
|
||||||
Name: fmt.Sprintf("%s_call_latency_microseconds", DefaultMetricPrefix),
|
|
||||||
Help: "Service client request latencies in microseconds",
|
|
||||||
},
|
|
||||||
[]string{"method"},
|
|
||||||
)
|
|
||||||
|
|
||||||
timeCounterHistogram := prometheus.NewHistogramVec(
|
|
||||||
prometheus.HistogramOpts{
|
|
||||||
Name: fmt.Sprintf("%s_call_request_duration_seconds", DefaultMetricPrefix),
|
|
||||||
Help: "Service client request time in seconds",
|
|
||||||
},
|
|
||||||
[]string{"method"},
|
|
||||||
)
|
|
||||||
|
|
||||||
for _, collector := range []prometheus.Collector{opsCounter, timeCounterSummary, timeCounterHistogram} {
|
|
||||||
if err := prometheus.DefaultRegisterer.Register(collector); err != nil {
|
|
||||||
// if already registered, skip fatal
|
|
||||||
if are, ok := err.(prometheus.AlreadyRegisteredError); ok {
|
|
||||||
collector = are.ExistingCollector
|
|
||||||
} else {
|
|
||||||
logger.Fatal(err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
handler := &wrapper{
|
handler := &wrapper{
|
||||||
timeCounterHistogram: timeCounterHistogram,
|
options: options,
|
||||||
timeCounterSummary: timeCounterSummary,
|
|
||||||
opsCounter: opsCounter,
|
|
||||||
callFunc: fn,
|
callFunc: fn,
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -147,20 +155,20 @@ func NewCallWrapper() client.CallWrapper {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (w *wrapper) CallFunc(ctx context.Context, node *registry.Node, req client.Request, rsp interface{}, opts client.CallOptions) error {
|
func (w *wrapper) CallFunc(ctx context.Context, node *registry.Node, req client.Request, rsp interface{}, opts client.CallOptions) error {
|
||||||
name := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint())
|
endpoint := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint())
|
||||||
|
|
||||||
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
|
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
|
||||||
us := v * 1000000 // make microseconds
|
us := v * 1000000 // make microseconds
|
||||||
w.timeCounterSummary.WithLabelValues(name).Observe(us)
|
timeCounterSummary.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(us)
|
||||||
w.timeCounterHistogram.WithLabelValues(name).Observe(v)
|
timeCounterHistogram.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(v)
|
||||||
}))
|
}))
|
||||||
defer timer.ObserveDuration()
|
defer timer.ObserveDuration()
|
||||||
|
|
||||||
err := w.callFunc(ctx, node, req, rsp, opts)
|
err := w.callFunc(ctx, node, req, rsp, opts)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
w.opsCounter.WithLabelValues(name, "success").Inc()
|
opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "success").Inc()
|
||||||
} else {
|
} else {
|
||||||
w.opsCounter.WithLabelValues(name, "fail").Inc()
|
opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "failure").Inc()
|
||||||
}
|
}
|
||||||
|
|
||||||
return err
|
return err
|
||||||
@ -168,110 +176,75 @@ func (w *wrapper) CallFunc(ctx context.Context, node *registry.Node, req client.
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (w *wrapper) Call(ctx context.Context, req client.Request, rsp interface{}, opts ...client.CallOption) error {
|
func (w *wrapper) Call(ctx context.Context, req client.Request, rsp interface{}, opts ...client.CallOption) error {
|
||||||
name := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint())
|
endpoint := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint())
|
||||||
|
|
||||||
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
|
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
|
||||||
us := v * 1000000 // make microseconds
|
us := v * 1000000 // make microseconds
|
||||||
w.timeCounterSummary.WithLabelValues(name).Observe(us)
|
timeCounterSummary.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(us)
|
||||||
w.timeCounterHistogram.WithLabelValues(name).Observe(v)
|
timeCounterHistogram.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(v)
|
||||||
}))
|
}))
|
||||||
defer timer.ObserveDuration()
|
defer timer.ObserveDuration()
|
||||||
|
|
||||||
err := w.Client.Call(ctx, req, rsp, opts...)
|
err := w.Client.Call(ctx, req, rsp, opts...)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
w.opsCounter.WithLabelValues(name, "success").Inc()
|
opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "success").Inc()
|
||||||
} else {
|
} else {
|
||||||
w.opsCounter.WithLabelValues(name, "fail").Inc()
|
opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "failure").Inc()
|
||||||
}
|
}
|
||||||
|
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
func (w *wrapper) Stream(ctx context.Context, req client.Request, opts ...client.CallOption) (client.Stream, error) {
|
func (w *wrapper) Stream(ctx context.Context, req client.Request, opts ...client.CallOption) (client.Stream, error) {
|
||||||
name := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint())
|
endpoint := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint())
|
||||||
|
|
||||||
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
|
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
|
||||||
us := v * 1000000 // make microseconds
|
us := v * 1000000 // make microseconds
|
||||||
w.timeCounterSummary.WithLabelValues(name).Observe(us)
|
timeCounterSummary.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(us)
|
||||||
w.timeCounterHistogram.WithLabelValues(name).Observe(v)
|
timeCounterHistogram.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(v)
|
||||||
}))
|
}))
|
||||||
defer timer.ObserveDuration()
|
defer timer.ObserveDuration()
|
||||||
|
|
||||||
stream, err := w.Client.Stream(ctx, req, opts...)
|
stream, err := w.Client.Stream(ctx, req, opts...)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
w.opsCounter.WithLabelValues(name, "success").Inc()
|
opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "success").Inc()
|
||||||
} else {
|
} else {
|
||||||
w.opsCounter.WithLabelValues(name, "fail").Inc()
|
opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "failure").Inc()
|
||||||
}
|
}
|
||||||
|
|
||||||
return stream, err
|
return stream, err
|
||||||
}
|
}
|
||||||
|
|
||||||
func (w *wrapper) Publish(ctx context.Context, p client.Message, opts ...client.PublishOption) error {
|
func (w *wrapper) Publish(ctx context.Context, p client.Message, opts ...client.PublishOption) error {
|
||||||
name := p.Topic()
|
endpoint := p.Topic()
|
||||||
|
|
||||||
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
|
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
|
||||||
us := v * 1000000 // make microseconds
|
us := v * 1000000 // make microseconds
|
||||||
w.timeCounterSummary.WithLabelValues(name).Observe(us)
|
timeCounterSummary.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(us)
|
||||||
w.timeCounterHistogram.WithLabelValues(name).Observe(v)
|
timeCounterHistogram.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(v)
|
||||||
}))
|
}))
|
||||||
defer timer.ObserveDuration()
|
defer timer.ObserveDuration()
|
||||||
|
|
||||||
err := w.Client.Publish(ctx, p, opts...)
|
err := w.Client.Publish(ctx, p, opts...)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
w.opsCounter.WithLabelValues(name, "success").Inc()
|
opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "success").Inc()
|
||||||
} else {
|
} else {
|
||||||
w.opsCounter.WithLabelValues(name, "fail").Inc()
|
opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "failure").Inc()
|
||||||
}
|
}
|
||||||
|
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewHandlerWrapper(opts ...server.Option) server.HandlerWrapper {
|
func NewHandlerWrapper(opts ...Option) server.HandlerWrapper {
|
||||||
labels := getLabels(opts...)
|
registerMetrics()
|
||||||
|
|
||||||
opsCounter := prometheus.NewCounterVec(
|
options := Options{}
|
||||||
prometheus.CounterOpts{
|
for _, opt := range opts {
|
||||||
ConstLabels: labels,
|
opt(&options)
|
||||||
Name: fmt.Sprintf("%s_handler_request_total", DefaultMetricPrefix),
|
|
||||||
Help: "How many requests processed, partitioned by method and status",
|
|
||||||
},
|
|
||||||
[]string{"method", "status"},
|
|
||||||
)
|
|
||||||
|
|
||||||
timeCounterSummary := prometheus.NewSummaryVec(
|
|
||||||
prometheus.SummaryOpts{
|
|
||||||
ConstLabels: labels,
|
|
||||||
Name: fmt.Sprintf("%s_handler_latency_microseconds", DefaultMetricPrefix),
|
|
||||||
Help: "Service handler request latencies in microseconds",
|
|
||||||
},
|
|
||||||
[]string{"method"},
|
|
||||||
)
|
|
||||||
|
|
||||||
timeCounterHistogram := prometheus.NewHistogramVec(
|
|
||||||
prometheus.HistogramOpts{
|
|
||||||
ConstLabels: labels,
|
|
||||||
Name: fmt.Sprintf("%s_handler_request_duration_seconds", DefaultMetricPrefix),
|
|
||||||
Help: "Service handler request time in seconds",
|
|
||||||
},
|
|
||||||
[]string{"method"},
|
|
||||||
)
|
|
||||||
|
|
||||||
for _, collector := range []prometheus.Collector{opsCounter, timeCounterSummary, timeCounterHistogram} {
|
|
||||||
if err := prometheus.DefaultRegisterer.Register(collector); err != nil {
|
|
||||||
// if already registered, skip fatal
|
|
||||||
if are, ok := err.(prometheus.AlreadyRegisteredError); ok {
|
|
||||||
collector = are.ExistingCollector
|
|
||||||
} else {
|
|
||||||
logger.Fatal(err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
handler := &wrapper{
|
handler := &wrapper{
|
||||||
timeCounterHistogram: timeCounterHistogram,
|
options: options,
|
||||||
timeCounterSummary: timeCounterSummary,
|
|
||||||
opsCounter: opsCounter,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return handler.HandlerFunc
|
return handler.HandlerFunc
|
||||||
@ -279,71 +252,36 @@ func NewHandlerWrapper(opts ...server.Option) server.HandlerWrapper {
|
|||||||
|
|
||||||
func (w *wrapper) HandlerFunc(fn server.HandlerFunc) server.HandlerFunc {
|
func (w *wrapper) HandlerFunc(fn server.HandlerFunc) server.HandlerFunc {
|
||||||
return func(ctx context.Context, req server.Request, rsp interface{}) error {
|
return func(ctx context.Context, req server.Request, rsp interface{}) error {
|
||||||
name := req.Endpoint()
|
endpoint := req.Endpoint()
|
||||||
|
|
||||||
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
|
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
|
||||||
us := v * 1000000 // make microseconds
|
us := v * 1000000 // make microseconds
|
||||||
w.timeCounterSummary.WithLabelValues(name).Observe(us)
|
timeCounterSummary.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(us)
|
||||||
w.timeCounterHistogram.WithLabelValues(name).Observe(v)
|
timeCounterHistogram.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(v)
|
||||||
}))
|
}))
|
||||||
defer timer.ObserveDuration()
|
defer timer.ObserveDuration()
|
||||||
|
|
||||||
err := fn(ctx, req, rsp)
|
err := fn(ctx, req, rsp)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
w.opsCounter.WithLabelValues(name, "success").Inc()
|
opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "success").Inc()
|
||||||
} else {
|
} else {
|
||||||
w.opsCounter.WithLabelValues(name, "fail").Inc()
|
opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "failure").Inc()
|
||||||
}
|
}
|
||||||
|
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewSubscriberWrapper(opts ...server.Option) server.SubscriberWrapper {
|
func NewSubscriberWrapper(opts ...Option) server.SubscriberWrapper {
|
||||||
labels := getLabels(opts...)
|
registerMetrics()
|
||||||
|
|
||||||
opsCounter := prometheus.NewCounterVec(
|
options := Options{}
|
||||||
prometheus.CounterOpts{
|
for _, opt := range opts {
|
||||||
ConstLabels: labels,
|
opt(&options)
|
||||||
Name: fmt.Sprintf("%s_subscriber_request_total", DefaultMetricPrefix),
|
|
||||||
Help: "How many requests processed, partitioned by topic and status",
|
|
||||||
},
|
|
||||||
[]string{"topic", "status"},
|
|
||||||
)
|
|
||||||
|
|
||||||
timeCounterSummary := prometheus.NewSummaryVec(
|
|
||||||
prometheus.SummaryOpts{
|
|
||||||
ConstLabels: labels,
|
|
||||||
Name: fmt.Sprintf("%s_subscriber_latency_microseconds", DefaultMetricPrefix),
|
|
||||||
Help: "Service subscriber request latencies in microseconds",
|
|
||||||
},
|
|
||||||
[]string{"topic"},
|
|
||||||
)
|
|
||||||
|
|
||||||
timeCounterHistogram := prometheus.NewHistogramVec(
|
|
||||||
prometheus.HistogramOpts{
|
|
||||||
ConstLabels: labels,
|
|
||||||
Name: fmt.Sprintf("%s_subscriber_request_duration_seconds", DefaultMetricPrefix),
|
|
||||||
Help: "Service subscriber request time in seconds",
|
|
||||||
},
|
|
||||||
[]string{"topic"},
|
|
||||||
)
|
|
||||||
|
|
||||||
for _, collector := range []prometheus.Collector{opsCounter, timeCounterSummary, timeCounterHistogram} {
|
|
||||||
if err := prometheus.DefaultRegisterer.Register(collector); err != nil {
|
|
||||||
// if already registered, skip fatal
|
|
||||||
if are, ok := err.(prometheus.AlreadyRegisteredError); ok {
|
|
||||||
collector = are.ExistingCollector
|
|
||||||
} else {
|
|
||||||
logger.Fatal(err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
handler := &wrapper{
|
handler := &wrapper{
|
||||||
timeCounterHistogram: timeCounterHistogram,
|
options: options,
|
||||||
timeCounterSummary: timeCounterSummary,
|
|
||||||
opsCounter: opsCounter,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return handler.SubscriberFunc
|
return handler.SubscriberFunc
|
||||||
@ -351,20 +289,20 @@ func NewSubscriberWrapper(opts ...server.Option) server.SubscriberWrapper {
|
|||||||
|
|
||||||
func (w *wrapper) SubscriberFunc(fn server.SubscriberFunc) server.SubscriberFunc {
|
func (w *wrapper) SubscriberFunc(fn server.SubscriberFunc) server.SubscriberFunc {
|
||||||
return func(ctx context.Context, msg server.Message) error {
|
return func(ctx context.Context, msg server.Message) error {
|
||||||
topic := msg.Topic()
|
endpoint := msg.Topic()
|
||||||
|
|
||||||
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
|
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
|
||||||
us := v * 1000000 // make microseconds
|
us := v * 1000000 // make microseconds
|
||||||
w.timeCounterSummary.WithLabelValues(topic).Observe(us)
|
timeCounterSummary.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(us)
|
||||||
w.timeCounterHistogram.WithLabelValues(topic).Observe(v)
|
timeCounterHistogram.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(v)
|
||||||
}))
|
}))
|
||||||
defer timer.ObserveDuration()
|
defer timer.ObserveDuration()
|
||||||
|
|
||||||
err := fn(ctx, msg)
|
err := fn(ctx, msg)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
w.opsCounter.WithLabelValues(topic, "success").Inc()
|
opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "success").Inc()
|
||||||
} else {
|
} else {
|
||||||
w.opsCounter.WithLabelValues(topic, "fail").Inc()
|
opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "failure").Inc()
|
||||||
}
|
}
|
||||||
|
|
||||||
return err
|
return err
|
||||||
|
@ -45,10 +45,6 @@ func TestPrometheusMetrics(t *testing.T) {
|
|||||||
id := "id-1234567890"
|
id := "id-1234567890"
|
||||||
version := "1.2.3.4"
|
version := "1.2.3.4"
|
||||||
|
|
||||||
md := make(map[string]string)
|
|
||||||
md["dc"] = "dc1"
|
|
||||||
md["node"] = "node1"
|
|
||||||
|
|
||||||
c := client.NewClient(client.Selector(sel))
|
c := client.NewClient(client.Selector(sel))
|
||||||
s := server.NewServer(
|
s := server.NewServer(
|
||||||
server.Name(name),
|
server.Name(name),
|
||||||
@ -58,10 +54,9 @@ func TestPrometheusMetrics(t *testing.T) {
|
|||||||
server.Broker(brk),
|
server.Broker(brk),
|
||||||
server.WrapHandler(
|
server.WrapHandler(
|
||||||
promwrapper.NewHandlerWrapper(
|
promwrapper.NewHandlerWrapper(
|
||||||
server.Metadata(md),
|
promwrapper.ServiceName(name),
|
||||||
server.Name(name),
|
promwrapper.ServiceVersion(version),
|
||||||
server.Version(version),
|
promwrapper.ServiceID(id),
|
||||||
server.Id(id),
|
|
||||||
),
|
),
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
@ -90,7 +85,7 @@ func TestPrometheusMetrics(t *testing.T) {
|
|||||||
|
|
||||||
list, _ := prometheus.DefaultGatherer.Gather()
|
list, _ := prometheus.DefaultGatherer.Gather()
|
||||||
|
|
||||||
metric := findMetricByName(list, dto.MetricType_SUMMARY, "micro_handler_latency_microseconds")
|
metric := findMetricByName(list, dto.MetricType_SUMMARY, "micro_latency_microseconds")
|
||||||
|
|
||||||
if metric == nil || metric.Metric == nil || len(metric.Metric) == 0 {
|
if metric == nil || metric.Metric == nil || len(metric.Metric) == 0 {
|
||||||
t.Fatalf("no metrics returned")
|
t.Fatalf("no metrics returned")
|
||||||
@ -98,17 +93,13 @@ func TestPrometheusMetrics(t *testing.T) {
|
|||||||
|
|
||||||
for _, v := range metric.Metric[0].Label {
|
for _, v := range metric.Metric[0].Label {
|
||||||
switch *v.Name {
|
switch *v.Name {
|
||||||
case "micro_dc":
|
|
||||||
assert.Equal(t, "dc1", *v.Value)
|
|
||||||
case "micro_node":
|
|
||||||
assert.Equal(t, "node1", *v.Value)
|
|
||||||
case "micro_version":
|
case "micro_version":
|
||||||
assert.Equal(t, version, *v.Value)
|
assert.Equal(t, version, *v.Value)
|
||||||
case "micro_id":
|
case "micro_id":
|
||||||
assert.Equal(t, id, *v.Value)
|
assert.Equal(t, id, *v.Value)
|
||||||
case "micro_name":
|
case "micro_name":
|
||||||
assert.Equal(t, name, *v.Value)
|
assert.Equal(t, name, *v.Value)
|
||||||
case "method":
|
case "micro_endpoint":
|
||||||
assert.Equal(t, "Test.Method", *v.Value)
|
assert.Equal(t, "Test.Method", *v.Value)
|
||||||
default:
|
default:
|
||||||
t.Fatalf("unknown %v with %v", *v.Name, *v.Value)
|
t.Fatalf("unknown %v with %v", *v.Name, *v.Value)
|
||||||
@ -118,21 +109,17 @@ func TestPrometheusMetrics(t *testing.T) {
|
|||||||
assert.Equal(t, uint64(2), *metric.Metric[0].Summary.SampleCount)
|
assert.Equal(t, uint64(2), *metric.Metric[0].Summary.SampleCount)
|
||||||
assert.True(t, *metric.Metric[0].Summary.SampleSum > 0)
|
assert.True(t, *metric.Metric[0].Summary.SampleSum > 0)
|
||||||
|
|
||||||
metric = findMetricByName(list, dto.MetricType_HISTOGRAM, "micro_handler_request_duration_seconds")
|
metric = findMetricByName(list, dto.MetricType_HISTOGRAM, "micro_request_duration_seconds")
|
||||||
|
|
||||||
for _, v := range metric.Metric[0].Label {
|
for _, v := range metric.Metric[0].Label {
|
||||||
switch *v.Name {
|
switch *v.Name {
|
||||||
case "micro_dc":
|
|
||||||
assert.Equal(t, "dc1", *v.Value)
|
|
||||||
case "micro_node":
|
|
||||||
assert.Equal(t, "node1", *v.Value)
|
|
||||||
case "micro_version":
|
case "micro_version":
|
||||||
assert.Equal(t, version, *v.Value)
|
assert.Equal(t, version, *v.Value)
|
||||||
case "micro_id":
|
case "micro_id":
|
||||||
assert.Equal(t, id, *v.Value)
|
assert.Equal(t, id, *v.Value)
|
||||||
case "micro_name":
|
case "micro_name":
|
||||||
assert.Equal(t, name, *v.Value)
|
assert.Equal(t, name, *v.Value)
|
||||||
case "method":
|
case "micro_endpoint":
|
||||||
assert.Equal(t, "Test.Method", *v.Value)
|
assert.Equal(t, "Test.Method", *v.Value)
|
||||||
default:
|
default:
|
||||||
t.Fatalf("unknown %v with %v", *v.Name, *v.Value)
|
t.Fatalf("unknown %v with %v", *v.Name, *v.Value)
|
||||||
@ -142,43 +129,35 @@ func TestPrometheusMetrics(t *testing.T) {
|
|||||||
assert.Equal(t, uint64(2), *metric.Metric[0].Histogram.SampleCount)
|
assert.Equal(t, uint64(2), *metric.Metric[0].Histogram.SampleCount)
|
||||||
assert.True(t, *metric.Metric[0].Histogram.SampleSum > 0)
|
assert.True(t, *metric.Metric[0].Histogram.SampleSum > 0)
|
||||||
|
|
||||||
metric = findMetricByName(list, dto.MetricType_COUNTER, "micro_handler_request_total")
|
metric = findMetricByName(list, dto.MetricType_COUNTER, "micro_request_total")
|
||||||
|
|
||||||
for _, v := range metric.Metric[0].Label {
|
for _, v := range metric.Metric[0].Label {
|
||||||
switch *v.Name {
|
switch *v.Name {
|
||||||
case "micro_dc":
|
|
||||||
assert.Equal(t, "dc1", *v.Value)
|
|
||||||
case "micro_node":
|
|
||||||
assert.Equal(t, "node1", *v.Value)
|
|
||||||
case "micro_version":
|
case "micro_version":
|
||||||
assert.Equal(t, version, *v.Value)
|
assert.Equal(t, version, *v.Value)
|
||||||
case "micro_id":
|
case "micro_id":
|
||||||
assert.Equal(t, id, *v.Value)
|
assert.Equal(t, id, *v.Value)
|
||||||
case "micro_name":
|
case "micro_name":
|
||||||
assert.Equal(t, name, *v.Value)
|
assert.Equal(t, name, *v.Value)
|
||||||
case "method":
|
case "micro_endpoint":
|
||||||
assert.Equal(t, "Test.Method", *v.Value)
|
assert.Equal(t, "Test.Method", *v.Value)
|
||||||
case "status":
|
case "micro_status":
|
||||||
assert.Equal(t, "fail", *v.Value)
|
assert.Equal(t, "failure", *v.Value)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
assert.Equal(t, *metric.Metric[0].Counter.Value, float64(1))
|
assert.Equal(t, *metric.Metric[0].Counter.Value, float64(1))
|
||||||
|
|
||||||
for _, v := range metric.Metric[1].Label {
|
for _, v := range metric.Metric[1].Label {
|
||||||
switch *v.Name {
|
switch *v.Name {
|
||||||
case "dc":
|
|
||||||
assert.Equal(t, "dc1", *v.Value)
|
|
||||||
case "node":
|
|
||||||
assert.Equal(t, "node1", *v.Value)
|
|
||||||
case "micro_version":
|
case "micro_version":
|
||||||
assert.Equal(t, version, *v.Value)
|
assert.Equal(t, version, *v.Value)
|
||||||
case "micro_id":
|
case "micro_id":
|
||||||
assert.Equal(t, id, *v.Value)
|
assert.Equal(t, id, *v.Value)
|
||||||
case "micro_name":
|
case "micro_name":
|
||||||
assert.Equal(t, name, *v.Value)
|
assert.Equal(t, name, *v.Value)
|
||||||
case "method":
|
case "micro_endpoint":
|
||||||
assert.Equal(t, "Test.Method", *v.Value)
|
assert.Equal(t, "Test.Method", *v.Value)
|
||||||
case "status":
|
case "micro_status":
|
||||||
assert.Equal(t, "success", *v.Value)
|
assert.Equal(t, "success", *v.Value)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user