wrapper/monitoring/prometheus: fix multiservice wrapper usage

Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org>
This commit is contained in:
Василий Толстов 2020-06-10 16:04:12 +03:00
parent 9cb28538ec
commit f2c41beb07
2 changed files with 168 additions and 251 deletions

View File

@ -3,6 +3,7 @@ package prometheus
import ( import (
"context" "context"
"fmt" "fmt"
"sync"
"github.com/micro/go-micro/v2/client" "github.com/micro/go-micro/v2/client"
"github.com/micro/go-micro/v2/logger" "github.com/micro/go-micro/v2/logger"
@ -13,84 +14,121 @@ import (
var ( var (
// default metric prefix // default metric prefix
DefaultMetricPrefix = "micro" DefaultMetricPrefix = "micro_"
// default label prefix // default label prefix
DefaultLabelPrefix = "micro" DefaultLabelPrefix = "micro_"
)
type wrapper struct {
opsCounter *prometheus.CounterVec opsCounter *prometheus.CounterVec
timeCounterSummary *prometheus.SummaryVec timeCounterSummary *prometheus.SummaryVec
timeCounterHistogram *prometheus.HistogramVec timeCounterHistogram *prometheus.HistogramVec
callFunc client.CallFunc
client.Client mu sync.Mutex
)
type Options struct {
Name string
Version string
ID string
} }
func getLabels(opts ...server.Option) prometheus.Labels { type Option func(*Options)
sopts := server.Options{}
for _, opt := range opts { func ServiceName(name string) Option {
opt(&sopts) return func(opts *Options) {
opts.Name = name
} }
labels := make(prometheus.Labels, len(sopts.Metadata))
for k, v := range sopts.Metadata {
labels[fmt.Sprintf("%s_%s", DefaultLabelPrefix, k)] = v
}
if len(sopts.Name) > 0 {
labels[fmt.Sprintf("%s_%s", DefaultLabelPrefix, "name")] = sopts.Name
}
if len(sopts.Id) > 0 {
labels[fmt.Sprintf("%s_%s", DefaultLabelPrefix, "id")] = sopts.Id
}
if len(sopts.Version) > 0 {
labels[fmt.Sprintf("%s_%s", DefaultLabelPrefix, "version")] = sopts.Version
}
return labels
} }
func NewClientWrapper() client.Wrapper { func ServiceVersion(version string) Option {
return func(c client.Client) client.Client { return func(opts *Options) {
opsCounter := prometheus.NewCounterVec( opts.Version = version
}
}
func ServiceID(id string) Option {
return func(opts *Options) {
opts.ID = id
}
}
func registerMetrics() {
mu.Lock()
defer mu.Unlock()
if opsCounter == nil {
opsCounter = prometheus.NewCounterVec(
prometheus.CounterOpts{ prometheus.CounterOpts{
Name: fmt.Sprintf("%s_client_request_total", DefaultMetricPrefix), Name: fmt.Sprintf("%srequest_total", DefaultMetricPrefix),
Help: "How many requests called, partitioned by method and status", Help: "Requests processed, partitioned by endpoint and status",
},
[]string{
fmt.Sprintf("%s%s", DefaultLabelPrefix, "name"),
fmt.Sprintf("%s%s", DefaultLabelPrefix, "version"),
fmt.Sprintf("%s%s", DefaultLabelPrefix, "id"),
fmt.Sprintf("%s%s", DefaultLabelPrefix, "endpoint"),
fmt.Sprintf("%s%s", DefaultLabelPrefix, "status"),
}, },
[]string{"method", "status"},
) )
}
timeCounterSummary := prometheus.NewSummaryVec( if timeCounterSummary == nil {
timeCounterSummary = prometheus.NewSummaryVec(
prometheus.SummaryOpts{ prometheus.SummaryOpts{
Name: fmt.Sprintf("%s_client_latency_microseconds", DefaultMetricPrefix), Name: fmt.Sprintf("%slatency_microseconds", DefaultMetricPrefix),
Help: "Service client request latencies in microseconds", Help: "Request latencies in microseconds, partitioned by endpoint",
},
[]string{
fmt.Sprintf("%s%s", DefaultLabelPrefix, "name"),
fmt.Sprintf("%s%s", DefaultLabelPrefix, "version"),
fmt.Sprintf("%s%s", DefaultLabelPrefix, "id"),
fmt.Sprintf("%s%s", DefaultLabelPrefix, "endpoint"),
}, },
[]string{"method"},
) )
}
timeCounterHistogram := prometheus.NewHistogramVec( if timeCounterHistogram == nil {
timeCounterHistogram = prometheus.NewHistogramVec(
prometheus.HistogramOpts{ prometheus.HistogramOpts{
Name: fmt.Sprintf("%s_client_request_duration_seconds", DefaultMetricPrefix), Name: fmt.Sprintf("%srequest_duration_seconds", DefaultMetricPrefix),
Help: "Service client request time in seconds", Help: "Request time in seconds, partitioned by endpoint",
},
[]string{
fmt.Sprintf("%s%s", DefaultLabelPrefix, "name"),
fmt.Sprintf("%s%s", DefaultLabelPrefix, "version"),
fmt.Sprintf("%s%s", DefaultLabelPrefix, "id"),
fmt.Sprintf("%s%s", DefaultLabelPrefix, "endpoint"),
}, },
[]string{"method"},
) )
}
for _, collector := range []prometheus.Collector{opsCounter, timeCounterSummary, timeCounterHistogram} { for _, collector := range []prometheus.Collector{opsCounter, timeCounterSummary, timeCounterHistogram} {
if err := prometheus.DefaultRegisterer.Register(collector); err != nil { if err := prometheus.DefaultRegisterer.Register(collector); err != nil {
// if already registered, skip fatal // if already registered, skip fatal
if are, ok := err.(prometheus.AlreadyRegisteredError); ok { if _, ok := err.(prometheus.AlreadyRegisteredError); !ok {
collector = are.ExistingCollector
} else {
logger.Fatal(err) logger.Fatal(err)
} }
} }
} }
}
type wrapper struct {
options Options
callFunc client.CallFunc
client.Client
}
func NewClientWrapper(opts ...Option) client.Wrapper {
registerMetrics()
options := Options{}
for _, opt := range opts {
opt(&options)
}
return func(c client.Client) client.Client {
handler := &wrapper{ handler := &wrapper{
timeCounterHistogram: timeCounterHistogram, options: options,
timeCounterSummary: timeCounterSummary,
opsCounter: opsCounter,
Client: c, Client: c,
} }
@ -98,47 +136,17 @@ func NewClientWrapper() client.Wrapper {
} }
} }
func NewCallWrapper() client.CallWrapper { func NewCallWrapper(opts ...Option) client.CallWrapper {
registerMetrics()
options := Options{}
for _, opt := range opts {
opt(&options)
}
return func(fn client.CallFunc) client.CallFunc { return func(fn client.CallFunc) client.CallFunc {
opsCounter := prometheus.NewCounterVec(
prometheus.CounterOpts{
Name: fmt.Sprintf("%s_call_request_total", DefaultMetricPrefix),
Help: "How many requests called, partitioned by method and status",
},
[]string{"method", "status"},
)
timeCounterSummary := prometheus.NewSummaryVec(
prometheus.SummaryOpts{
Name: fmt.Sprintf("%s_call_latency_microseconds", DefaultMetricPrefix),
Help: "Service client request latencies in microseconds",
},
[]string{"method"},
)
timeCounterHistogram := prometheus.NewHistogramVec(
prometheus.HistogramOpts{
Name: fmt.Sprintf("%s_call_request_duration_seconds", DefaultMetricPrefix),
Help: "Service client request time in seconds",
},
[]string{"method"},
)
for _, collector := range []prometheus.Collector{opsCounter, timeCounterSummary, timeCounterHistogram} {
if err := prometheus.DefaultRegisterer.Register(collector); err != nil {
// if already registered, skip fatal
if are, ok := err.(prometheus.AlreadyRegisteredError); ok {
collector = are.ExistingCollector
} else {
logger.Fatal(err)
}
}
}
handler := &wrapper{ handler := &wrapper{
timeCounterHistogram: timeCounterHistogram, options: options,
timeCounterSummary: timeCounterSummary,
opsCounter: opsCounter,
callFunc: fn, callFunc: fn,
} }
@ -147,20 +155,20 @@ func NewCallWrapper() client.CallWrapper {
} }
func (w *wrapper) CallFunc(ctx context.Context, node *registry.Node, req client.Request, rsp interface{}, opts client.CallOptions) error { func (w *wrapper) CallFunc(ctx context.Context, node *registry.Node, req client.Request, rsp interface{}, opts client.CallOptions) error {
name := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint()) endpoint := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint())
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) { timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
us := v * 1000000 // make microseconds us := v * 1000000 // make microseconds
w.timeCounterSummary.WithLabelValues(name).Observe(us) timeCounterSummary.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(us)
w.timeCounterHistogram.WithLabelValues(name).Observe(v) timeCounterHistogram.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(v)
})) }))
defer timer.ObserveDuration() defer timer.ObserveDuration()
err := w.callFunc(ctx, node, req, rsp, opts) err := w.callFunc(ctx, node, req, rsp, opts)
if err == nil { if err == nil {
w.opsCounter.WithLabelValues(name, "success").Inc() opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "success").Inc()
} else { } else {
w.opsCounter.WithLabelValues(name, "fail").Inc() opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "failure").Inc()
} }
return err return err
@ -168,110 +176,75 @@ func (w *wrapper) CallFunc(ctx context.Context, node *registry.Node, req client.
} }
func (w *wrapper) Call(ctx context.Context, req client.Request, rsp interface{}, opts ...client.CallOption) error { func (w *wrapper) Call(ctx context.Context, req client.Request, rsp interface{}, opts ...client.CallOption) error {
name := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint()) endpoint := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint())
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) { timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
us := v * 1000000 // make microseconds us := v * 1000000 // make microseconds
w.timeCounterSummary.WithLabelValues(name).Observe(us) timeCounterSummary.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(us)
w.timeCounterHistogram.WithLabelValues(name).Observe(v) timeCounterHistogram.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(v)
})) }))
defer timer.ObserveDuration() defer timer.ObserveDuration()
err := w.Client.Call(ctx, req, rsp, opts...) err := w.Client.Call(ctx, req, rsp, opts...)
if err == nil { if err == nil {
w.opsCounter.WithLabelValues(name, "success").Inc() opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "success").Inc()
} else { } else {
w.opsCounter.WithLabelValues(name, "fail").Inc() opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "failure").Inc()
} }
return err return err
} }
func (w *wrapper) Stream(ctx context.Context, req client.Request, opts ...client.CallOption) (client.Stream, error) { func (w *wrapper) Stream(ctx context.Context, req client.Request, opts ...client.CallOption) (client.Stream, error) {
name := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint()) endpoint := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint())
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) { timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
us := v * 1000000 // make microseconds us := v * 1000000 // make microseconds
w.timeCounterSummary.WithLabelValues(name).Observe(us) timeCounterSummary.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(us)
w.timeCounterHistogram.WithLabelValues(name).Observe(v) timeCounterHistogram.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(v)
})) }))
defer timer.ObserveDuration() defer timer.ObserveDuration()
stream, err := w.Client.Stream(ctx, req, opts...) stream, err := w.Client.Stream(ctx, req, opts...)
if err == nil { if err == nil {
w.opsCounter.WithLabelValues(name, "success").Inc() opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "success").Inc()
} else { } else {
w.opsCounter.WithLabelValues(name, "fail").Inc() opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "failure").Inc()
} }
return stream, err return stream, err
} }
func (w *wrapper) Publish(ctx context.Context, p client.Message, opts ...client.PublishOption) error { func (w *wrapper) Publish(ctx context.Context, p client.Message, opts ...client.PublishOption) error {
name := p.Topic() endpoint := p.Topic()
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) { timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
us := v * 1000000 // make microseconds us := v * 1000000 // make microseconds
w.timeCounterSummary.WithLabelValues(name).Observe(us) timeCounterSummary.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(us)
w.timeCounterHistogram.WithLabelValues(name).Observe(v) timeCounterHistogram.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(v)
})) }))
defer timer.ObserveDuration() defer timer.ObserveDuration()
err := w.Client.Publish(ctx, p, opts...) err := w.Client.Publish(ctx, p, opts...)
if err == nil { if err == nil {
w.opsCounter.WithLabelValues(name, "success").Inc() opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "success").Inc()
} else { } else {
w.opsCounter.WithLabelValues(name, "fail").Inc() opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "failure").Inc()
} }
return err return err
} }
func NewHandlerWrapper(opts ...server.Option) server.HandlerWrapper { func NewHandlerWrapper(opts ...Option) server.HandlerWrapper {
labels := getLabels(opts...) registerMetrics()
opsCounter := prometheus.NewCounterVec( options := Options{}
prometheus.CounterOpts{ for _, opt := range opts {
ConstLabels: labels, opt(&options)
Name: fmt.Sprintf("%s_handler_request_total", DefaultMetricPrefix),
Help: "How many requests processed, partitioned by method and status",
},
[]string{"method", "status"},
)
timeCounterSummary := prometheus.NewSummaryVec(
prometheus.SummaryOpts{
ConstLabels: labels,
Name: fmt.Sprintf("%s_handler_latency_microseconds", DefaultMetricPrefix),
Help: "Service handler request latencies in microseconds",
},
[]string{"method"},
)
timeCounterHistogram := prometheus.NewHistogramVec(
prometheus.HistogramOpts{
ConstLabels: labels,
Name: fmt.Sprintf("%s_handler_request_duration_seconds", DefaultMetricPrefix),
Help: "Service handler request time in seconds",
},
[]string{"method"},
)
for _, collector := range []prometheus.Collector{opsCounter, timeCounterSummary, timeCounterHistogram} {
if err := prometheus.DefaultRegisterer.Register(collector); err != nil {
// if already registered, skip fatal
if are, ok := err.(prometheus.AlreadyRegisteredError); ok {
collector = are.ExistingCollector
} else {
logger.Fatal(err)
}
}
} }
handler := &wrapper{ handler := &wrapper{
timeCounterHistogram: timeCounterHistogram, options: options,
timeCounterSummary: timeCounterSummary,
opsCounter: opsCounter,
} }
return handler.HandlerFunc return handler.HandlerFunc
@ -279,71 +252,36 @@ func NewHandlerWrapper(opts ...server.Option) server.HandlerWrapper {
func (w *wrapper) HandlerFunc(fn server.HandlerFunc) server.HandlerFunc { func (w *wrapper) HandlerFunc(fn server.HandlerFunc) server.HandlerFunc {
return func(ctx context.Context, req server.Request, rsp interface{}) error { return func(ctx context.Context, req server.Request, rsp interface{}) error {
name := req.Endpoint() endpoint := req.Endpoint()
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) { timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
us := v * 1000000 // make microseconds us := v * 1000000 // make microseconds
w.timeCounterSummary.WithLabelValues(name).Observe(us) timeCounterSummary.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(us)
w.timeCounterHistogram.WithLabelValues(name).Observe(v) timeCounterHistogram.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(v)
})) }))
defer timer.ObserveDuration() defer timer.ObserveDuration()
err := fn(ctx, req, rsp) err := fn(ctx, req, rsp)
if err == nil { if err == nil {
w.opsCounter.WithLabelValues(name, "success").Inc() opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "success").Inc()
} else { } else {
w.opsCounter.WithLabelValues(name, "fail").Inc() opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "failure").Inc()
} }
return err return err
} }
} }
func NewSubscriberWrapper(opts ...server.Option) server.SubscriberWrapper { func NewSubscriberWrapper(opts ...Option) server.SubscriberWrapper {
labels := getLabels(opts...) registerMetrics()
opsCounter := prometheus.NewCounterVec( options := Options{}
prometheus.CounterOpts{ for _, opt := range opts {
ConstLabels: labels, opt(&options)
Name: fmt.Sprintf("%s_subscriber_request_total", DefaultMetricPrefix),
Help: "How many requests processed, partitioned by topic and status",
},
[]string{"topic", "status"},
)
timeCounterSummary := prometheus.NewSummaryVec(
prometheus.SummaryOpts{
ConstLabels: labels,
Name: fmt.Sprintf("%s_subscriber_latency_microseconds", DefaultMetricPrefix),
Help: "Service subscriber request latencies in microseconds",
},
[]string{"topic"},
)
timeCounterHistogram := prometheus.NewHistogramVec(
prometheus.HistogramOpts{
ConstLabels: labels,
Name: fmt.Sprintf("%s_subscriber_request_duration_seconds", DefaultMetricPrefix),
Help: "Service subscriber request time in seconds",
},
[]string{"topic"},
)
for _, collector := range []prometheus.Collector{opsCounter, timeCounterSummary, timeCounterHistogram} {
if err := prometheus.DefaultRegisterer.Register(collector); err != nil {
// if already registered, skip fatal
if are, ok := err.(prometheus.AlreadyRegisteredError); ok {
collector = are.ExistingCollector
} else {
logger.Fatal(err)
}
}
} }
handler := &wrapper{ handler := &wrapper{
timeCounterHistogram: timeCounterHistogram, options: options,
timeCounterSummary: timeCounterSummary,
opsCounter: opsCounter,
} }
return handler.SubscriberFunc return handler.SubscriberFunc
@ -351,20 +289,20 @@ func NewSubscriberWrapper(opts ...server.Option) server.SubscriberWrapper {
func (w *wrapper) SubscriberFunc(fn server.SubscriberFunc) server.SubscriberFunc { func (w *wrapper) SubscriberFunc(fn server.SubscriberFunc) server.SubscriberFunc {
return func(ctx context.Context, msg server.Message) error { return func(ctx context.Context, msg server.Message) error {
topic := msg.Topic() endpoint := msg.Topic()
timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) { timer := prometheus.NewTimer(prometheus.ObserverFunc(func(v float64) {
us := v * 1000000 // make microseconds us := v * 1000000 // make microseconds
w.timeCounterSummary.WithLabelValues(topic).Observe(us) timeCounterSummary.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(us)
w.timeCounterHistogram.WithLabelValues(topic).Observe(v) timeCounterHistogram.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint).Observe(v)
})) }))
defer timer.ObserveDuration() defer timer.ObserveDuration()
err := fn(ctx, msg) err := fn(ctx, msg)
if err == nil { if err == nil {
w.opsCounter.WithLabelValues(topic, "success").Inc() opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "success").Inc()
} else { } else {
w.opsCounter.WithLabelValues(topic, "fail").Inc() opsCounter.WithLabelValues(w.options.Name, w.options.Version, w.options.ID, endpoint, "failure").Inc()
} }
return err return err

View File

@ -45,10 +45,6 @@ func TestPrometheusMetrics(t *testing.T) {
id := "id-1234567890" id := "id-1234567890"
version := "1.2.3.4" version := "1.2.3.4"
md := make(map[string]string)
md["dc"] = "dc1"
md["node"] = "node1"
c := client.NewClient(client.Selector(sel)) c := client.NewClient(client.Selector(sel))
s := server.NewServer( s := server.NewServer(
server.Name(name), server.Name(name),
@ -58,10 +54,9 @@ func TestPrometheusMetrics(t *testing.T) {
server.Broker(brk), server.Broker(brk),
server.WrapHandler( server.WrapHandler(
promwrapper.NewHandlerWrapper( promwrapper.NewHandlerWrapper(
server.Metadata(md), promwrapper.ServiceName(name),
server.Name(name), promwrapper.ServiceVersion(version),
server.Version(version), promwrapper.ServiceID(id),
server.Id(id),
), ),
), ),
) )
@ -90,7 +85,7 @@ func TestPrometheusMetrics(t *testing.T) {
list, _ := prometheus.DefaultGatherer.Gather() list, _ := prometheus.DefaultGatherer.Gather()
metric := findMetricByName(list, dto.MetricType_SUMMARY, "micro_handler_latency_microseconds") metric := findMetricByName(list, dto.MetricType_SUMMARY, "micro_latency_microseconds")
if metric == nil || metric.Metric == nil || len(metric.Metric) == 0 { if metric == nil || metric.Metric == nil || len(metric.Metric) == 0 {
t.Fatalf("no metrics returned") t.Fatalf("no metrics returned")
@ -98,17 +93,13 @@ func TestPrometheusMetrics(t *testing.T) {
for _, v := range metric.Metric[0].Label { for _, v := range metric.Metric[0].Label {
switch *v.Name { switch *v.Name {
case "micro_dc":
assert.Equal(t, "dc1", *v.Value)
case "micro_node":
assert.Equal(t, "node1", *v.Value)
case "micro_version": case "micro_version":
assert.Equal(t, version, *v.Value) assert.Equal(t, version, *v.Value)
case "micro_id": case "micro_id":
assert.Equal(t, id, *v.Value) assert.Equal(t, id, *v.Value)
case "micro_name": case "micro_name":
assert.Equal(t, name, *v.Value) assert.Equal(t, name, *v.Value)
case "method": case "micro_endpoint":
assert.Equal(t, "Test.Method", *v.Value) assert.Equal(t, "Test.Method", *v.Value)
default: default:
t.Fatalf("unknown %v with %v", *v.Name, *v.Value) t.Fatalf("unknown %v with %v", *v.Name, *v.Value)
@ -118,21 +109,17 @@ func TestPrometheusMetrics(t *testing.T) {
assert.Equal(t, uint64(2), *metric.Metric[0].Summary.SampleCount) assert.Equal(t, uint64(2), *metric.Metric[0].Summary.SampleCount)
assert.True(t, *metric.Metric[0].Summary.SampleSum > 0) assert.True(t, *metric.Metric[0].Summary.SampleSum > 0)
metric = findMetricByName(list, dto.MetricType_HISTOGRAM, "micro_handler_request_duration_seconds") metric = findMetricByName(list, dto.MetricType_HISTOGRAM, "micro_request_duration_seconds")
for _, v := range metric.Metric[0].Label { for _, v := range metric.Metric[0].Label {
switch *v.Name { switch *v.Name {
case "micro_dc":
assert.Equal(t, "dc1", *v.Value)
case "micro_node":
assert.Equal(t, "node1", *v.Value)
case "micro_version": case "micro_version":
assert.Equal(t, version, *v.Value) assert.Equal(t, version, *v.Value)
case "micro_id": case "micro_id":
assert.Equal(t, id, *v.Value) assert.Equal(t, id, *v.Value)
case "micro_name": case "micro_name":
assert.Equal(t, name, *v.Value) assert.Equal(t, name, *v.Value)
case "method": case "micro_endpoint":
assert.Equal(t, "Test.Method", *v.Value) assert.Equal(t, "Test.Method", *v.Value)
default: default:
t.Fatalf("unknown %v with %v", *v.Name, *v.Value) t.Fatalf("unknown %v with %v", *v.Name, *v.Value)
@ -142,43 +129,35 @@ func TestPrometheusMetrics(t *testing.T) {
assert.Equal(t, uint64(2), *metric.Metric[0].Histogram.SampleCount) assert.Equal(t, uint64(2), *metric.Metric[0].Histogram.SampleCount)
assert.True(t, *metric.Metric[0].Histogram.SampleSum > 0) assert.True(t, *metric.Metric[0].Histogram.SampleSum > 0)
metric = findMetricByName(list, dto.MetricType_COUNTER, "micro_handler_request_total") metric = findMetricByName(list, dto.MetricType_COUNTER, "micro_request_total")
for _, v := range metric.Metric[0].Label { for _, v := range metric.Metric[0].Label {
switch *v.Name { switch *v.Name {
case "micro_dc":
assert.Equal(t, "dc1", *v.Value)
case "micro_node":
assert.Equal(t, "node1", *v.Value)
case "micro_version": case "micro_version":
assert.Equal(t, version, *v.Value) assert.Equal(t, version, *v.Value)
case "micro_id": case "micro_id":
assert.Equal(t, id, *v.Value) assert.Equal(t, id, *v.Value)
case "micro_name": case "micro_name":
assert.Equal(t, name, *v.Value) assert.Equal(t, name, *v.Value)
case "method": case "micro_endpoint":
assert.Equal(t, "Test.Method", *v.Value) assert.Equal(t, "Test.Method", *v.Value)
case "status": case "micro_status":
assert.Equal(t, "fail", *v.Value) assert.Equal(t, "failure", *v.Value)
} }
} }
assert.Equal(t, *metric.Metric[0].Counter.Value, float64(1)) assert.Equal(t, *metric.Metric[0].Counter.Value, float64(1))
for _, v := range metric.Metric[1].Label { for _, v := range metric.Metric[1].Label {
switch *v.Name { switch *v.Name {
case "dc":
assert.Equal(t, "dc1", *v.Value)
case "node":
assert.Equal(t, "node1", *v.Value)
case "micro_version": case "micro_version":
assert.Equal(t, version, *v.Value) assert.Equal(t, version, *v.Value)
case "micro_id": case "micro_id":
assert.Equal(t, id, *v.Value) assert.Equal(t, id, *v.Value)
case "micro_name": case "micro_name":
assert.Equal(t, name, *v.Value) assert.Equal(t, name, *v.Value)
case "method": case "micro_endpoint":
assert.Equal(t, "Test.Method", *v.Value) assert.Equal(t, "Test.Method", *v.Value)
case "status": case "micro_status":
assert.Equal(t, "success", *v.Value) assert.Equal(t, "success", *v.Value)
} }
} }