Compare commits

...

16 Commits
v3.10.11 ... v3

Author SHA1 Message Date
c59902201d add_labels (#120)
Some checks failed
codeql / analyze (go) (push) Failing after 56s
build / test (push) Failing after 4m59s
build / lint (push) Successful in 9m34s
closes #119

Reviewed-on: #120
Co-authored-by: Evstigneev Denis <danteevstigneev@yandex.ru>
Co-committed-by: Evstigneev Denis <danteevstigneev@yandex.ru>
2024-11-19 00:34:04 +03:00
0981f89f60 Merge pull request 'fix deps' (#118) from devstigneev/micro-store-redis:v3_1 into v3
Some checks failed
codeql / analyze (go) (push) Failing after 46s
build / test (push) Failing after 4m56s
build / lint (push) Successful in 9m31s
Reviewed-on: #118
2024-10-15 11:56:58 +03:00
332fe5f4d4 fix deps
Some checks failed
automerge / automerge (pull_request) Has been skipped
autoapprove / autoapprove (pull_request) Successful in 8s
dependabot-automerge / automerge (pull_request) Has been skipped
codeql / analyze (go) (pull_request) Failing after 49s
prbuild / test (pull_request) Has been cancelled
prbuild / lint (pull_request) Has been cancelled
2024-10-14 18:56:37 +03:00
757fe0245b update all
Some checks failed
codeql / analyze (go) (push) Failing after 42s
build / test (push) Failing after 4m53s
build / lint (push) Successful in 9m28s
Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org>
2024-10-14 17:49:38 +03:00
27eccc1ed2 fixup nil pointer assign
Some checks failed
codeql / analyze (go) (push) Failing after 41s
build / test (push) Failing after 4m52s
build / lint (push) Successful in 9m29s
Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org>
2024-10-08 16:52:26 +03:00
7c641fa8ac fixup nil pointer assign
Some checks failed
codeql / analyze (go) (push) Failing after 39s
build / test (push) Failing after 4m55s
build / lint (push) Has been cancelled
Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org>
2024-10-08 16:47:27 +03:00
24c9f20196 toolchain go1.22.4 (#116)
Some checks failed
codeql / analyze (go) (push) Failing after 44s
build / test (push) Failing after 4m55s
build / lint (push) Successful in 9m29s
- go mod tidy
- fix version redis
- retract (
  v9.5.3 // This version was accidentally released.
) from github.com/redis/go-redis/extra/rediscmd/v9@v9.5.3

Reviewed-on: #116
Co-authored-by: Evstigneev Denis <danteevstigneev@yandex.ru>
Co-committed-by: Evstigneev Denis <danteevstigneev@yandex.ru>
2024-10-08 13:07:46 +03:00
953b5b0021 fixup default redis default local addr
Some checks failed
codeql / analyze (go) (push) Failing after 37s
build / test (push) Failing after 4m54s
build / lint (push) Successful in 9m29s
Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org>
2024-10-06 18:49:50 +03:00
87e2e2b947 switch to universal client
Some checks failed
codeql / analyze (go) (push) Failing after 1m30s
build / test (push) Failing after 5m13s
build / lint (push) Successful in 9m40s
Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org>
2024-10-05 16:11:46 +03:00
256e61a437 fixup slice override
Some checks failed
build / test (push) Failing after 7s
build / lint (push) Failing after 7s
codeql / analyze (go) (push) Failing after 10s
Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org>
2024-09-25 15:27:53 +03:00
f9cdd41c94 Merge pull request 'changed MetricPrefix to Labels' (#112) from devstigneev/micro-store-redis:v3 into v3
Some checks failed
build / test (push) Failing after 10s
build / lint (push) Failing after 10s
codeql / analyze (go) (push) Failing after 10s
Reviewed-on: #112
Reviewed-by: Василий Толстов <v.tolstov@unistack.org>
2024-09-23 18:23:27 +03:00
ecad15fe17 changed MetricPrefix to Labels
Some checks failed
automerge / automerge (pull_request) Has been skipped
autoapprove / autoapprove (pull_request) Failing after 5s
dependabot-automerge / automerge (pull_request) Has been skipped
codeql / analyze (go) (pull_request) Has been cancelled
prbuild / test (pull_request) Has been cancelled
prbuild / lint (pull_request) Has been cancelled
2024-09-21 14:01:04 +03:00
fa3d18b353 tracing commented redis.dial
Some checks failed
build / test (push) Failing after 21s
build / lint (push) Successful in 20s
codeql / analyze (go) (push) Failing after 39s
Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org>
2024-07-18 11:28:45 +03:00
2f3951773f cleanup trace spans
Some checks failed
build / lint (push) Successful in 24s
build / test (push) Failing after 1m25s
codeql / analyze (go) (push) Failing after 1m53s
Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org>
2024-07-05 23:11:12 +03:00
b263e14032 cleanup trace spans from cluster slots command
Some checks failed
build / lint (push) Successful in 22s
build / test (push) Failing after 1m30s
codeql / analyze (go) (push) Failing after 1m51s
Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org>
2024-07-05 13:44:29 +03:00
518cc1db73 fixup duplicate nested redis span
Some checks failed
build / lint (push) Successful in 22s
build / test (push) Failing after 1m29s
codeql / analyze (go) (push) Failing after 1m52s
Signed-off-by: Vasiliy Tolstov <v.tolstov@unistack.org>
2024-07-05 12:59:05 +03:00
8 changed files with 322 additions and 272 deletions

1
.gitignore vendored Normal file
View File

@ -0,0 +1 @@
.idea

13
go.mod
View File

@ -1,16 +1,19 @@
module go.unistack.org/micro-store-redis/v3 module go.unistack.org/micro-store-redis/v3
go 1.21 go 1.22
toolchain go1.22.4 toolchain go1.22.4
require ( require (
github.com/redis/go-redis/extra/rediscmd/v9 v9.5.3 github.com/redis/go-redis/extra/rediscmd/v9 v9.6.2
github.com/redis/go-redis/v9 v9.5.3 github.com/redis/go-redis/v9 v9.6.2
go.unistack.org/micro/v3 v3.10.80 go.unistack.org/micro/v3 v3.10.97
) )
require ( require (
github.com/cespare/xxhash/v2 v2.2.0 // indirect github.com/cespare/xxhash/v2 v2.3.0 // indirect
github.com/dgryski/go-rendezvous v0.0.0-20200823014737-9f7001d12a5f // indirect github.com/dgryski/go-rendezvous v0.0.0-20200823014737-9f7001d12a5f // indirect
github.com/google/go-cmp v0.6.0 // indirect
go.unistack.org/micro-proto/v3 v3.4.1 // indirect
google.golang.org/protobuf v1.35.1 // indirect
) )

22
go.sum
View File

@ -2,13 +2,19 @@ github.com/bsm/ginkgo/v2 v2.12.0 h1:Ny8MWAHyOepLGlLKYmXG4IEkioBysk6GpaRTLC8zwWs=
github.com/bsm/ginkgo/v2 v2.12.0/go.mod h1:SwYbGRRDovPVboqFv0tPTcG1sN61LM1Z4ARdbAV9g4c= github.com/bsm/ginkgo/v2 v2.12.0/go.mod h1:SwYbGRRDovPVboqFv0tPTcG1sN61LM1Z4ARdbAV9g4c=
github.com/bsm/gomega v1.27.10 h1:yeMWxP2pV2fG3FgAODIY8EiRE3dy0aeFYt4l7wh6yKA= github.com/bsm/gomega v1.27.10 h1:yeMWxP2pV2fG3FgAODIY8EiRE3dy0aeFYt4l7wh6yKA=
github.com/bsm/gomega v1.27.10/go.mod h1:JyEr/xRbxbtgWNi8tIEVPUYZ5Dzef52k01W3YH0H+O0= github.com/bsm/gomega v1.27.10/go.mod h1:JyEr/xRbxbtgWNi8tIEVPUYZ5Dzef52k01W3YH0H+O0=
github.com/cespare/xxhash/v2 v2.2.0 h1:DC2CZ1Ep5Y4k3ZQ899DldepgrayRUGE6BBZ/cd9Cj44= github.com/cespare/xxhash/v2 v2.3.0 h1:UL815xU9SqsFlibzuggzjXhog7bL6oX9BbNZnL2UFvs=
github.com/cespare/xxhash/v2 v2.2.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= github.com/cespare/xxhash/v2 v2.3.0/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs=
github.com/dgryski/go-rendezvous v0.0.0-20200823014737-9f7001d12a5f h1:lO4WD4F/rVNCu3HqELle0jiPLLBs70cWOduZpkS1E78= github.com/dgryski/go-rendezvous v0.0.0-20200823014737-9f7001d12a5f h1:lO4WD4F/rVNCu3HqELle0jiPLLBs70cWOduZpkS1E78=
github.com/dgryski/go-rendezvous v0.0.0-20200823014737-9f7001d12a5f/go.mod h1:cuUVRXasLTGF7a8hSLbxyZXjz+1KgoB3wDUb6vlszIc= github.com/dgryski/go-rendezvous v0.0.0-20200823014737-9f7001d12a5f/go.mod h1:cuUVRXasLTGF7a8hSLbxyZXjz+1KgoB3wDUb6vlszIc=
github.com/redis/go-redis/extra/rediscmd/v9 v9.5.3 h1:1/BDligzCa40GTllkDnY3Y5DTHuKCONbB2JcRyIfl20= github.com/google/go-cmp v0.6.0 h1:ofyhxvXcZhMsU5ulbFiLKl/XBFqE1GSq7atu8tAmTRI=
github.com/redis/go-redis/extra/rediscmd/v9 v9.5.3/go.mod h1:3dZmcLn3Qw6FLlWASn1g4y+YO9ycEFUOM+bhBmzLVKQ= github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeNGIjoY=
github.com/redis/go-redis/v9 v9.5.3 h1:fOAp1/uJG+ZtcITgZOfYFmTKPE7n4Vclj1wZFgRciUU= github.com/redis/go-redis/extra/rediscmd/v9 v9.6.2 h1:oBlErygFka9FAVfowzW7sRUfB7n31YX0aFzfaOflM3w=
github.com/redis/go-redis/v9 v9.5.3/go.mod h1:hdY0cQFCN4fnSYT6TkisLufl/4W5UIXyv0b/CLO2V2M= github.com/redis/go-redis/extra/rediscmd/v9 v9.6.2/go.mod h1:jYLUE5tC8UsFnpSclhEjZlFLMMtHH4jTfFMnshNWkoo=
go.unistack.org/micro/v3 v3.10.80 h1:A0zWNoM9MOcMg9gdFFgVkgbT3uSYVIINhuvumX9nP2o= github.com/redis/go-redis/v9 v9.6.2 h1:w0uvkRbc9KpgD98zcvo5IrVUsn0lXpRMuhNgiHDJzdk=
go.unistack.org/micro/v3 v3.10.80/go.mod h1:erMgt3Bl7vQQ0e9UpQyR5NlLiZ9pKeEJ9+1tfYFaqUg= github.com/redis/go-redis/v9 v9.6.2/go.mod h1:0C0c6ycQsdpVNQpxb1njEQIqkx5UcsM8FJCQLgE9+RA=
go.unistack.org/micro-proto/v3 v3.4.1 h1:UTjLSRz2YZuaHk9iSlVqqsA50JQNAEK2ZFboGqtEa9Q=
go.unistack.org/micro-proto/v3 v3.4.1/go.mod h1:okx/cnOhzuCX0ggl/vToatbCupi0O44diiiLLsZ93Zo=
go.unistack.org/micro/v3 v3.10.97 h1:8l7fv+i06/PjPrBBhRC/ZQkWGIOuHPg3jJN0vktYE78=
go.unistack.org/micro/v3 v3.10.97/go.mod h1:YzMldzHN9Ei+zy5t/Psu7RUWDZwUfrNYiStSQtTz90g=
google.golang.org/protobuf v1.35.1 h1:m3LfL6/Ca+fqnjnlqQXNpFPABW1UD7mjh8KO2mKFytA=
google.golang.org/protobuf v1.35.1/go.mod h1:9fA7Ob0pmnwhb644+1+CVWFRbNajQ6iRojtC/QF5bRE=

View File

@ -1,9 +1,7 @@
package redis package redis
import ( import (
"time" goredis "github.com/redis/go-redis/v9"
"github.com/redis/go-redis/v9"
"go.unistack.org/micro/v3/logger" "go.unistack.org/micro/v3/logger"
"go.unistack.org/micro/v3/meter" "go.unistack.org/micro/v3/meter"
"go.unistack.org/micro/v3/store" "go.unistack.org/micro/v3/store"
@ -12,30 +10,34 @@ import (
type configKey struct{} type configKey struct{}
func Config(c *redis.Options) store.Option { func Config(c *goredis.Options) store.Option {
return store.SetOption(configKey{}, c) return store.SetOption(configKey{}, c)
} }
type clusterConfigKey struct{} type clusterConfigKey struct{}
func ClusterConfig(c *redis.ClusterOptions) store.Option { func ClusterConfig(c *goredis.ClusterOptions) store.Option {
return store.SetOption(clusterConfigKey{}, c) return store.SetOption(clusterConfigKey{}, c)
} }
type universalConfigKey struct{}
func UniversalConfig(c *goredis.UniversalOptions) store.Option {
return store.SetOption(universalConfigKey{}, c)
}
var ( var (
// DefaultMeterStatsInterval holds default stats interval labelHost = "redis_host"
DefaultMeterStatsInterval = 5 * time.Second labelName = "redis_name"
// DefaultMeterMetricPrefix holds default metric prefix
DefaultMeterMetricPrefix = "micro_store_"
) )
// Options struct holds wrapper options // Options struct holds wrapper options
type Options struct { type Options struct {
Logger logger.Logger Logger logger.Logger
Meter meter.Meter Meter meter.Meter
Tracer tracer.Tracer Tracer tracer.Tracer
MeterMetricPrefix string RedisHost string
MeterStatsInterval time.Duration RedisName string
} }
// Option func signature // Option func signature
@ -44,11 +46,9 @@ type Option func(*Options)
// NewOptions create new Options struct from provided option slice // NewOptions create new Options struct from provided option slice
func NewOptions(opts ...Option) Options { func NewOptions(opts ...Option) Options {
options := Options{ options := Options{
Logger: logger.DefaultLogger, Logger: logger.DefaultLogger,
Meter: meter.DefaultMeter, Meter: meter.DefaultMeter,
Tracer: tracer.DefaultTracer, Tracer: tracer.DefaultTracer,
MeterStatsInterval: DefaultMeterStatsInterval,
MeterMetricPrefix: DefaultMeterMetricPrefix,
} }
for _, o := range opts { for _, o := range opts {
@ -56,24 +56,12 @@ func NewOptions(opts ...Option) Options {
} }
options.Meter = options.Meter.Clone( options.Meter = options.Meter.Clone(
meter.MetricPrefix(options.MeterMetricPrefix), meter.Labels(
labelHost, options.RedisHost,
labelName, options.RedisName),
) )
options.Logger = options.Logger.Clone(logger.WithCallerSkipCount(1)) options.Logger = options.Logger.Clone(logger.WithAddCallerSkipCount(1))
return options return options
} }
// MetricInterval specifies stats interval for *sql.DB
func MetricInterval(td time.Duration) Option {
return func(o *Options) {
o.MeterStatsInterval = td
}
}
// MetricPrefix specifies prefix for each metric
func MetricPrefix(pref string) Option {
return func(o *Options) {
o.MeterMetricPrefix = pref
}
}

390
redis.go
View File

@ -2,12 +2,12 @@ package redis
import ( import (
"context" "context"
"fmt" "errors"
"reflect" "reflect"
"strings" "strings"
"time" "time"
redis "github.com/redis/go-redis/v9" goredis "github.com/redis/go-redis/v9"
"go.unistack.org/micro/v3/semconv" "go.unistack.org/micro/v3/semconv"
"go.unistack.org/micro/v3/store" "go.unistack.org/micro/v3/store"
pool "go.unistack.org/micro/v3/util/xpool" pool "go.unistack.org/micro/v3/util/xpool"
@ -16,7 +16,7 @@ import (
var ( var (
DefaultPathSeparator = "/" DefaultPathSeparator = "/"
DefaultClusterOptions = &redis.ClusterOptions{ DefaultUniversalOptions = &goredis.UniversalOptions{
Username: "", Username: "",
Password: "", // no password set Password: "", // no password set
MaxRetries: 2, MaxRetries: 2,
@ -28,7 +28,19 @@ var (
MinIdleConns: 10, MinIdleConns: 10,
} }
DefaultOptions = &redis.Options{ DefaultClusterOptions = &goredis.ClusterOptions{
Username: "",
Password: "", // no password set
MaxRetries: 2,
MaxRetryBackoff: 256 * time.Millisecond,
DialTimeout: 1 * time.Second,
ReadTimeout: 1 * time.Second,
WriteTimeout: 1 * time.Second,
PoolTimeout: 1 * time.Second,
MinIdleConns: 10,
}
DefaultOptions = &goredis.Options{
Username: "", Username: "",
Password: "", // no password set Password: "", // no password set
DB: 0, // use default DB DB: 0, // use default DB
@ -44,22 +56,16 @@ var (
type Store struct { type Store struct {
opts store.Options opts store.Options
cli *wrappedClient cli goredis.UniversalClient
done chan struct{} done chan struct{}
pool pool.Pool[*strings.Builder] pool *pool.StringsPool
}
type wrappedClient struct {
*redis.Client
*redis.ClusterClient
} }
func (r *Store) Connect(ctx context.Context) error { func (r *Store) Connect(ctx context.Context) error {
var err error if r.cli == nil {
if r.cli.Client != nil { return store.ErrNotConnected
err = r.cli.Client.Ping(ctx).Err()
} }
err = r.cli.ClusterClient.Ping(ctx).Err() err := r.cli.Ping(ctx).Err()
setSpanError(ctx, err) setSpanError(ctx, err)
return err return err
} }
@ -77,16 +83,20 @@ func (r *Store) Init(opts ...store.Option) error {
return nil return nil
} }
func (r *Store) Client() *redis.Client { func (r *Store) Client() *goredis.Client {
if r.cli.Client != nil { if c, ok := r.cli.(*goredis.Client); ok {
return r.cli.Client return c
} }
return nil return nil
} }
func (r *Store) ClusterClient() *redis.ClusterClient { func (r *Store) UniversalClient() goredis.UniversalClient {
if r.cli.ClusterClient != nil { return r.cli
return r.cli.ClusterClient }
func (r *Store) ClusterClient() *goredis.ClusterClient {
if c, ok := r.cli.(*goredis.ClusterClient); ok {
return c
} }
return nil return nil
} }
@ -97,10 +107,8 @@ func (r *Store) Disconnect(ctx context.Context) error {
case <-r.done: case <-r.done:
return err return err
default: default:
if r.cli.Client != nil { if r.cli != nil {
err = r.cli.Client.Close() err = r.cli.Close()
} else if r.cli.ClusterClient != nil {
err = r.cli.ClusterClient.Close()
} }
close(r.done) close(r.done)
return err return err
@ -108,7 +116,11 @@ func (r *Store) Disconnect(ctx context.Context) error {
} }
func (r *Store) Exists(ctx context.Context, key string, opts ...store.ExistsOption) error { func (r *Store) Exists(ctx context.Context, key string, opts ...store.ExistsOption) error {
b := r.pool.Get()
defer r.pool.Put(b)
options := store.NewExistsOptions(opts...) options := store.NewExistsOptions(opts...)
labels := make([]string, 0, 6)
labels = append(labels, "name", options.Name, "statement", "exists")
timeout := r.opts.Timeout timeout := r.opts.Timeout
if options.Timeout > 0 { if options.Timeout > 0 {
@ -121,29 +133,24 @@ func (r *Store) Exists(ctx context.Context, key string, opts ...store.ExistsOpti
defer cancel() defer cancel()
} }
rkey := r.getKey(r.opts.Namespace, options.Namespace, key) rkey := r.getKey(b, r.opts.Namespace, options.Namespace, key)
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Inc() r.opts.Meter.Counter(semconv.StoreRequestInflight, labels...).Inc()
ts := time.Now() ts := time.Now()
var err error val, err := r.cli.Exists(ctx, rkey).Result()
var val int64
if r.cli.Client != nil {
val, err = r.cli.Client.Exists(ctx, rkey).Result()
} else {
val, err = r.cli.ClusterClient.Exists(ctx, rkey).Result()
}
setSpanError(ctx, err) setSpanError(ctx, err)
te := time.Since(ts) te := time.Since(ts)
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Dec() r.opts.Meter.Counter(semconv.StoreRequestInflight, labels...).Dec()
r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, "name", options.Name).Update(te.Seconds())
r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, "name", options.Name).Update(te.Seconds()) r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, labels...).Update(te.Seconds())
if err == redis.Nil || (err == nil && val == 0) { r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, labels...).Update(te.Seconds())
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc() if errors.Is(err, goredis.Nil) || (err == nil && val == 0) {
r.opts.Meter.Counter(semconv.StoreRequestTotal, append(labels, "status", "miss")...).Inc()
return store.ErrNotFound return store.ErrNotFound
} else if err == nil { } else if err == nil {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "hit").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, append(labels, "status", "hit")...).Inc()
} else if err != nil { } else if err != nil {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "failure").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, append(labels, "status", "failure")...).Inc()
return err return err
} }
@ -151,7 +158,12 @@ func (r *Store) Exists(ctx context.Context, key string, opts ...store.ExistsOpti
} }
func (r *Store) Read(ctx context.Context, key string, val interface{}, opts ...store.ReadOption) error { func (r *Store) Read(ctx context.Context, key string, val interface{}, opts ...store.ReadOption) error {
b := r.pool.Get()
defer r.pool.Put(b)
options := store.NewReadOptions(opts...) options := store.NewReadOptions(opts...)
labels := make([]string, 0, 6)
labels = append(labels, "name", options.Name, "statement", "read")
timeout := r.opts.Timeout timeout := r.opts.Timeout
if options.Timeout > 0 { if options.Timeout > 0 {
@ -164,29 +176,23 @@ func (r *Store) Read(ctx context.Context, key string, val interface{}, opts ...s
defer cancel() defer cancel()
} }
rkey := r.getKey(r.opts.Namespace, options.Namespace, key) rkey := r.getKey(b, r.opts.Namespace, options.Namespace, key)
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Inc() r.opts.Meter.Counter(semconv.StoreRequestInflight, labels...).Inc()
ts := time.Now() ts := time.Now()
var buf []byte buf, err := r.cli.Get(ctx, rkey).Bytes()
var err error
if r.cli.Client != nil {
buf, err = r.cli.Client.Get(ctx, rkey).Bytes()
} else {
buf, err = r.cli.ClusterClient.Get(ctx, rkey).Bytes()
}
setSpanError(ctx, err) setSpanError(ctx, err)
te := time.Since(ts) te := time.Since(ts)
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Dec() r.opts.Meter.Counter(semconv.StoreRequestInflight, labels...).Dec()
r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, "name", options.Name).Update(te.Seconds()) r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, labels...).Update(te.Seconds())
r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, "name", options.Name).Update(te.Seconds()) r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, labels...).Update(te.Seconds())
if err == redis.Nil || (err == nil && buf == nil) { if errors.Is(err, goredis.Nil) || (err == nil && buf == nil) {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, append(labels, "status", "miss")...).Inc()
return store.ErrNotFound return store.ErrNotFound
} else if err == nil { } else if err == nil {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "hit").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, append(labels, "status", "hit")...).Inc()
} else if err != nil { } else if err != nil {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "failure").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, append(labels, "status", "failure")...).Inc()
return err return err
} }
@ -218,9 +224,15 @@ func (r *Store) MRead(ctx context.Context, keys []string, vals interface{}, opts
defer cancel() defer cancel()
} }
var rkeys []string
var pools []*strings.Builder
if r.opts.Namespace != "" || options.Namespace != "" { if r.opts.Namespace != "" || options.Namespace != "" {
rkeys = make([]string, len(keys))
pools = make([]*strings.Builder, len(keys))
for idx, key := range keys { for idx, key := range keys {
keys[idx] = r.getKey(r.opts.Namespace, options.Namespace, key) b := r.pool.Get()
pools[idx] = b
rkeys[idx] = r.getKey(b, r.opts.Namespace, options.Namespace, key)
} }
} }
@ -228,17 +240,20 @@ func (r *Store) MRead(ctx context.Context, keys []string, vals interface{}, opts
ts := time.Now() ts := time.Now()
var rvals []interface{} var rvals []interface{}
var err error var err error
if r.cli.Client != nil { if r.opts.Namespace != "" || options.Namespace != "" {
rvals, err = r.cli.Client.MGet(ctx, keys...).Result() rvals, err = r.cli.MGet(ctx, rkeys...).Result()
for idx := range pools {
r.pool.Put(pools[idx])
}
} else { } else {
rvals, err = r.cli.ClusterClient.MGet(ctx, keys...).Result() rvals, err = r.cli.MGet(ctx, keys...).Result()
} }
setSpanError(ctx, err) setSpanError(ctx, err)
te := time.Since(ts) te := time.Since(ts)
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Dec() r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Dec()
r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, "name", options.Name).Update(te.Seconds()) r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, "name", options.Name).Update(te.Seconds())
r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, "name", options.Name).Update(te.Seconds()) r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, "name", options.Name).Update(te.Seconds())
if err == redis.Nil || (len(rvals) == 0) { if err == goredis.Nil || (len(rvals) == 0) {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc()
return store.ErrNotFound return store.ErrNotFound
} else if err == nil { } else if err == nil {
@ -308,26 +323,35 @@ func (r *Store) MDelete(ctx context.Context, keys []string, opts ...store.Delete
defer cancel() defer cancel()
} }
var rkeys []string
var pools []*strings.Builder
if r.opts.Namespace != "" || options.Namespace != "" { if r.opts.Namespace != "" || options.Namespace != "" {
rkeys = make([]string, len(keys))
pools = make([]*strings.Builder, len(keys))
for idx, key := range keys { for idx, key := range keys {
keys[idx] = r.getKey(r.opts.Namespace, options.Namespace, key) b := r.pool.Get()
pools[idx] = b
rkeys[idx] = r.getKey(b, r.opts.Namespace, options.Namespace, key)
} }
} }
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Inc() r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Inc()
ts := time.Now() ts := time.Now()
var err error var err error
if r.cli.Client != nil { if r.opts.Namespace != "" || options.Namespace != "" {
err = r.cli.Client.Del(ctx, keys...).Err() err = r.cli.Del(ctx, rkeys...).Err()
for idx := range pools {
r.pool.Put(pools[idx])
}
} else { } else {
err = r.cli.ClusterClient.Del(ctx, keys...).Err() err = r.cli.Del(ctx, keys...).Err()
} }
setSpanError(ctx, err) setSpanError(ctx, err)
te := time.Since(ts) te := time.Since(ts)
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Dec() r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Dec()
r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, "name", options.Name).Update(te.Seconds()) r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, "name", options.Name).Update(te.Seconds())
r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, "name", options.Name).Update(te.Seconds()) r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, "name", options.Name).Update(te.Seconds())
if err == redis.Nil { if err == goredis.Nil {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc()
return store.ErrNotFound return store.ErrNotFound
} else if err == nil { } else if err == nil {
@ -341,7 +365,12 @@ func (r *Store) MDelete(ctx context.Context, keys []string, opts ...store.Delete
} }
func (r *Store) Delete(ctx context.Context, key string, opts ...store.DeleteOption) error { func (r *Store) Delete(ctx context.Context, key string, opts ...store.DeleteOption) error {
b := r.pool.Get()
defer r.pool.Put(b)
options := store.NewDeleteOptions(opts...) options := store.NewDeleteOptions(opts...)
labels := make([]string, 0, 6)
labels = append(labels, "name", options.Name, "statement", "delete")
timeout := r.opts.Timeout timeout := r.opts.Timeout
if options.Timeout > 0 { if options.Timeout > 0 {
@ -354,26 +383,22 @@ func (r *Store) Delete(ctx context.Context, key string, opts ...store.DeleteOpti
defer cancel() defer cancel()
} }
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Inc() r.opts.Meter.Counter(semconv.StoreRequestInflight, labels...).Inc()
ts := time.Now() ts := time.Now()
var err error err := r.cli.Del(ctx, r.getKey(b, r.opts.Namespace, options.Namespace, key)).Err()
if r.cli.Client != nil {
err = r.cli.Client.Del(ctx, r.getKey(r.opts.Namespace, options.Namespace, key)).Err()
} else {
err = r.cli.ClusterClient.Del(ctx, r.getKey(r.opts.Namespace, options.Namespace, key)).Err()
}
setSpanError(ctx, err)
te := time.Since(ts) te := time.Since(ts)
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Dec() setSpanError(ctx, err)
r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, "name", options.Name).Update(te.Seconds())
r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, "name", options.Name).Update(te.Seconds()) r.opts.Meter.Counter(semconv.StoreRequestInflight, labels...).Dec()
if err == redis.Nil { r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, labels...).Update(te.Seconds())
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc() r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, labels...).Update(te.Seconds())
if errors.Is(err, goredis.Nil) {
r.opts.Meter.Counter(semconv.StoreRequestTotal, append(labels, "status", "miss")...).Inc()
return store.ErrNotFound return store.ErrNotFound
} else if err == nil { } else if err == nil {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "hit").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, append(labels, "status", "hit")...).Inc()
} else if err != nil { } else if err != nil {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "failure").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, append(labels, "status", "failure")...).Inc()
return err return err
} }
@ -395,9 +420,11 @@ func (r *Store) MWrite(ctx context.Context, keys []string, vals []interface{}, o
} }
kvs := make([]string, 0, len(keys)*2) kvs := make([]string, 0, len(keys)*2)
pools := make([]*strings.Builder, len(keys))
for idx, key := range keys { for idx, key := range keys {
kvs = append(kvs, r.getKey(r.opts.Namespace, options.Namespace, key)) b := r.pool.Get()
pools[idx] = b
kvs = append(kvs, r.getKey(b, r.opts.Namespace, options.Namespace, key))
switch vt := vals[idx].(type) { switch vt := vals[idx].(type) {
case string: case string:
@ -414,9 +441,8 @@ func (r *Store) MWrite(ctx context.Context, keys []string, vals []interface{}, o
} }
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Inc() r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Inc()
ts := time.Now()
pipeliner := func(pipe redis.Pipeliner) error { pipeliner := func(pipe goredis.Pipeliner) error {
for idx := 0; idx < len(kvs); idx += 2 { for idx := 0; idx < len(kvs); idx += 2 {
if _, err := pipe.Set(ctx, kvs[idx], kvs[idx+1], options.TTL).Result(); err != nil { if _, err := pipe.Set(ctx, kvs[idx], kvs[idx+1], options.TTL).Result(); err != nil {
setSpanError(ctx, err) setSpanError(ctx, err)
@ -426,20 +452,18 @@ func (r *Store) MWrite(ctx context.Context, keys []string, vals []interface{}, o
return nil return nil
} }
var err error ts := time.Now()
var cmds []redis.Cmder cmds, err := r.cli.Pipelined(ctx, pipeliner)
for idx := range pools {
if r.cli.Client != nil { r.pool.Put(pools[idx])
cmds, err = r.cli.Client.Pipelined(ctx, pipeliner)
} else {
cmds, err = r.cli.ClusterClient.Pipelined(ctx, pipeliner)
} }
setSpanError(ctx, err)
te := time.Since(ts) te := time.Since(ts)
setSpanError(ctx, err)
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Dec() r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Dec()
r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, "name", options.Name).Update(te.Seconds()) r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, "name", options.Name).Update(te.Seconds())
r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, "name", options.Name).Update(te.Seconds()) r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, "name", options.Name).Update(te.Seconds())
if err == redis.Nil { if err == goredis.Nil {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc()
return store.ErrNotFound return store.ErrNotFound
} else if err == nil { } else if err == nil {
@ -451,7 +475,7 @@ func (r *Store) MWrite(ctx context.Context, keys []string, vals []interface{}, o
for _, cmd := range cmds { for _, cmd := range cmds {
if err = cmd.Err(); err != nil { if err = cmd.Err(); err != nil {
if err == redis.Nil { if err == goredis.Nil {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc()
return store.ErrNotFound return store.ErrNotFound
} }
@ -465,7 +489,12 @@ func (r *Store) MWrite(ctx context.Context, keys []string, vals []interface{}, o
} }
func (r *Store) Write(ctx context.Context, key string, val interface{}, opts ...store.WriteOption) error { func (r *Store) Write(ctx context.Context, key string, val interface{}, opts ...store.WriteOption) error {
b := r.pool.Get()
defer r.pool.Put(b)
options := store.NewWriteOptions(opts...) options := store.NewWriteOptions(opts...)
labels := make([]string, 0, 6)
labels = append(labels, "name", options.Name, "statement", "write")
timeout := r.opts.Timeout timeout := r.opts.Timeout
if options.Timeout > 0 { if options.Timeout > 0 {
@ -478,7 +507,7 @@ func (r *Store) Write(ctx context.Context, key string, val interface{}, opts ...
defer cancel() defer cancel()
} }
rkey := r.getKey(r.opts.Namespace, options.Namespace, key) rkey := r.getKey(b, r.opts.Namespace, options.Namespace, key)
var buf []byte var buf []byte
switch vt := val.(type) { switch vt := val.(type) {
@ -494,26 +523,22 @@ func (r *Store) Write(ctx context.Context, key string, val interface{}, opts ...
} }
} }
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Inc() r.opts.Meter.Counter(semconv.StoreRequestInflight, labels...).Inc()
ts := time.Now() ts := time.Now()
var err error err := r.cli.Set(ctx, rkey, buf, options.TTL).Err()
if r.cli.Client != nil {
err = r.cli.Client.Set(ctx, rkey, buf, options.TTL).Err()
} else {
err = r.cli.ClusterClient.Set(ctx, rkey, buf, options.TTL).Err()
}
setSpanError(ctx, err)
te := time.Since(ts) te := time.Since(ts)
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Dec() setSpanError(ctx, err)
r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, "name", options.Name).Update(te.Seconds())
r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, "name", options.Name).Update(te.Seconds()) r.opts.Meter.Counter(semconv.StoreRequestInflight, labels...).Dec()
if err == redis.Nil { r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, labels...).Update(te.Seconds())
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc() r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, labels...).Update(te.Seconds())
if errors.Is(err, goredis.Nil) {
r.opts.Meter.Counter(semconv.StoreRequestTotal, append(labels, "status", "miss")...).Inc()
return store.ErrNotFound return store.ErrNotFound
} else if err == nil { } else if err == nil {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "hit").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, append(labels, "status", "hit")...).Inc()
} else if err != nil { } else if err != nil {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "failure").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, append(labels, "status", "failure")...).Inc()
return err return err
} }
@ -521,12 +546,15 @@ func (r *Store) Write(ctx context.Context, key string, val interface{}, opts ...
} }
func (r *Store) List(ctx context.Context, opts ...store.ListOption) ([]string, error) { func (r *Store) List(ctx context.Context, opts ...store.ListOption) ([]string, error) {
b := r.pool.Get()
defer r.pool.Put(b)
options := store.NewListOptions(opts...) options := store.NewListOptions(opts...)
if len(options.Namespace) == 0 { if len(options.Namespace) == 0 {
options.Namespace = r.opts.Namespace options.Namespace = r.opts.Namespace
} }
rkey := r.getKey(options.Namespace, "", options.Prefix+"*") rkey := r.getKey(b, options.Namespace, "", options.Prefix+"*")
if options.Suffix != "" { if options.Suffix != "" {
rkey += options.Suffix rkey += options.Suffix
} }
@ -548,10 +576,8 @@ func (r *Store) List(ctx context.Context, opts ...store.ListOption) ([]string, e
var keys []string var keys []string
var err error var err error
if r.cli.Client != nil { if c, ok := r.cli.(*goredis.ClusterClient); ok {
keys, err = r.cli.Client.Keys(ctx, rkey).Result() err = c.ForEachMaster(ctx, func(nctx context.Context, cli *goredis.Client) error {
} else {
err = r.cli.ClusterClient.ForEachMaster(ctx, func(nctx context.Context, cli *redis.Client) error {
nkeys, nerr := cli.Keys(nctx, rkey).Result() nkeys, nerr := cli.Keys(nctx, rkey).Result()
if nerr != nil { if nerr != nil {
return nerr return nerr
@ -559,13 +585,16 @@ func (r *Store) List(ctx context.Context, opts ...store.ListOption) ([]string, e
keys = append(keys, nkeys...) keys = append(keys, nkeys...)
return nil return nil
}) })
} else {
keys, err = r.cli.Keys(ctx, rkey).Result()
} }
setSpanError(ctx, err)
te := time.Since(ts) te := time.Since(ts)
setSpanError(ctx, err)
r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Dec() r.opts.Meter.Counter(semconv.StoreRequestInflight, "name", options.Name).Dec()
r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, "name", options.Name).Update(te.Seconds()) r.opts.Meter.Summary(semconv.StoreRequestLatencyMicroseconds, "name", options.Name).Update(te.Seconds())
r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, "name", options.Name).Update(te.Seconds()) r.opts.Meter.Histogram(semconv.StoreRequestDurationSeconds, "name", options.Name).Update(te.Seconds())
if err == redis.Nil { if err == goredis.Nil {
r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc() r.opts.Meter.Counter(semconv.StoreRequestTotal, "name", options.Name, "status", "miss").Inc()
return nil, store.ErrNotFound return nil, store.ErrNotFound
} else if err == nil { } else if err == nil {
@ -607,79 +636,102 @@ func NewStore(opts ...store.Option) *Store {
} }
func (r *Store) configure() error { func (r *Store) configure() error {
var redisOptions *redis.Options
var redisClusterOptions *redis.ClusterOptions
var err error
nodes := r.opts.Addrs
if len(nodes) == 0 {
nodes = []string{"redis://127.0.0.1:6379"}
}
if r.cli != nil && r.opts.Context == nil { if r.cli != nil && r.opts.Context == nil {
return nil return nil
} }
universalOptions := DefaultUniversalOptions
if r.opts.Context != nil { if r.opts.Context != nil {
if c, ok := r.opts.Context.Value(configKey{}).(*redis.Options); ok { if o, ok := r.opts.Context.Value(configKey{}).(*goredis.Options); ok {
redisOptions = c universalOptions.Addrs = []string{o.Addr}
universalOptions.Dialer = o.Dialer
universalOptions.OnConnect = o.OnConnect
universalOptions.Username = o.Username
universalOptions.Password = o.Password
universalOptions.MaxRetries = o.MaxRetries
universalOptions.MinRetryBackoff = o.MinRetryBackoff
universalOptions.MaxRetryBackoff = o.MaxRetryBackoff
universalOptions.DialTimeout = o.DialTimeout
universalOptions.ReadTimeout = o.ReadTimeout
universalOptions.WriteTimeout = o.WriteTimeout
universalOptions.ContextTimeoutEnabled = o.ContextTimeoutEnabled
universalOptions.PoolFIFO = o.PoolFIFO
universalOptions.PoolSize = o.PoolSize
universalOptions.PoolTimeout = o.PoolTimeout
universalOptions.MinIdleConns = o.MinIdleConns
universalOptions.MaxIdleConns = o.MaxIdleConns
universalOptions.ConnMaxIdleTime = o.ConnMaxIdleTime
universalOptions.ConnMaxLifetime = o.ConnMaxLifetime
if r.opts.TLSConfig != nil { if r.opts.TLSConfig != nil {
redisOptions.TLSConfig = r.opts.TLSConfig universalOptions.TLSConfig = r.opts.TLSConfig
} }
} }
if c, ok := r.opts.Context.Value(clusterConfigKey{}).(*redis.ClusterOptions); ok { if o, ok := r.opts.Context.Value(clusterConfigKey{}).(*goredis.ClusterOptions); ok {
redisClusterOptions = c universalOptions.Addrs = o.Addrs
universalOptions.Dialer = o.Dialer
universalOptions.OnConnect = o.OnConnect
universalOptions.Username = o.Username
universalOptions.Password = o.Password
universalOptions.MaxRedirects = o.MaxRedirects
universalOptions.ReadOnly = o.ReadOnly
universalOptions.RouteByLatency = o.RouteByLatency
universalOptions.RouteRandomly = o.RouteRandomly
universalOptions.MaxRetries = o.MaxRetries
universalOptions.MinRetryBackoff = o.MinRetryBackoff
universalOptions.MaxRetryBackoff = o.MaxRetryBackoff
universalOptions.DialTimeout = o.DialTimeout
universalOptions.ReadTimeout = o.ReadTimeout
universalOptions.WriteTimeout = o.WriteTimeout
universalOptions.ContextTimeoutEnabled = o.ContextTimeoutEnabled
universalOptions.PoolFIFO = o.PoolFIFO
universalOptions.PoolSize = o.PoolSize
universalOptions.PoolTimeout = o.PoolTimeout
universalOptions.MinIdleConns = o.MinIdleConns
universalOptions.MaxIdleConns = o.MaxIdleConns
universalOptions.ConnMaxIdleTime = o.ConnMaxIdleTime
universalOptions.ConnMaxLifetime = o.ConnMaxLifetime
if r.opts.TLSConfig != nil { if r.opts.TLSConfig != nil {
redisClusterOptions.TLSConfig = r.opts.TLSConfig universalOptions.TLSConfig = r.opts.TLSConfig
}
}
if o, ok := r.opts.Context.Value(universalConfigKey{}).(*goredis.UniversalOptions); ok {
universalOptions = o
if r.opts.TLSConfig != nil {
universalOptions.TLSConfig = r.opts.TLSConfig
} }
} }
} }
if redisOptions != nil && redisClusterOptions != nil { if len(r.opts.Addrs) > 0 {
return fmt.Errorf("must specify only one option Config or ClusterConfig") universalOptions.Addrs = r.opts.Addrs
} else {
universalOptions.Addrs = []string{"127.0.0.1:6379"}
} }
if redisOptions == nil && redisClusterOptions == nil && r.cli != nil { r.cli = goredis.NewUniversalClient(universalOptions)
return nil setTracing(r.cli, r.opts.Tracer)
}
if redisOptions == nil && redisClusterOptions == nil && len(nodes) == 1 { r.pool = pool.NewStringsPool(50)
redisOptions, err = redis.ParseURL(nodes[0])
if err != nil {
redisOptions = DefaultOptions
redisOptions.Addr = r.opts.Addrs[0]
redisOptions.TLSConfig = r.opts.TLSConfig
}
} else if redisOptions == nil && redisClusterOptions == nil && len(nodes) > 1 {
redisClusterOptions = DefaultClusterOptions
redisClusterOptions.Addrs = r.opts.Addrs
redisClusterOptions.TLSConfig = r.opts.TLSConfig
}
if redisOptions != nil {
c := redis.NewClient(redisOptions)
setTracing(c, r.opts.Tracer)
r.cli = &wrappedClient{Client: c}
} else if redisClusterOptions != nil {
c := redis.NewClusterClient(redisClusterOptions)
setTracing(c, r.opts.Tracer)
r.cli = &wrappedClient{ClusterClient: c}
}
r.pool = pool.NewPool(func() *strings.Builder { return &strings.Builder{} })
r.statsMeter() r.statsMeter()
return nil return nil
} }
func (r *Store) getKey(mainNamespace string, opNamespace string, key string) string { func (r *Store) getKey(b *strings.Builder, mainNamespace string, opNamespace string, key string) string {
b := r.pool.Get()
defer r.pool.Put(b)
b.Reset()
if opNamespace == "" { if opNamespace == "" {
opNamespace = mainNamespace opNamespace = mainNamespace
} }

View File

@ -7,6 +7,7 @@ import (
"testing" "testing"
"time" "time"
goredis "github.com/redis/go-redis/v9"
"go.unistack.org/micro/v3/store" "go.unistack.org/micro/v3/store"
"go.unistack.org/micro/v3/tracer" "go.unistack.org/micro/v3/tracer"
) )
@ -41,7 +42,7 @@ func TestKeepTTL(t *testing.T) {
func Test_rkv_configure(t *testing.T) { func Test_rkv_configure(t *testing.T) {
type fields struct { type fields struct {
options store.Options options store.Options
Client *wrappedClient Client goredis.UniversalClient
} }
type wantValues struct { type wantValues struct {
username string username string

View File

@ -3,7 +3,8 @@ package redis
import ( import (
"time" "time"
"github.com/redis/go-redis/v9" goredis "github.com/redis/go-redis/v9"
"go.unistack.org/micro/v3/meter"
) )
var ( var (
@ -13,45 +14,36 @@ var (
PoolConnTotalCurrent = "pool_conn_total_current" PoolConnTotalCurrent = "pool_conn_total_current"
PoolConnIdleCurrent = "pool_conn_idle_current" PoolConnIdleCurrent = "pool_conn_idle_current"
PoolConnStaleTotal = "pool_conn_stale_total" PoolConnStaleTotal = "pool_conn_stale_total"
meterRequestTotal = "request_total"
meterRequestLatencyMicroseconds = "latency_microseconds"
meterRequestDurationSeconds = "request_duration_seconds"
) )
type Statser interface { type Statser interface {
PoolStats() *redis.PoolStats PoolStats() *goredis.PoolStats
} }
func (r *Store) statsMeter() { func (r *Store) statsMeter() {
var st Statser var st Statser
if r.cli.Client != nil { if r.cli != nil {
st = r.cli.Client st = r.cli
} else if r.cli.ClusterClient != nil {
st = r.cli.ClusterClient
} else { } else {
return return
} }
go func() { go func() {
ticker := time.NewTicker(DefaultMeterStatsInterval) ticker := time.NewTicker(meter.DefaultMeterStatsInterval)
defer ticker.Stop() defer ticker.Stop()
for { for _ = range ticker.C {
select { if st == nil {
case <-ticker.C: return
if st == nil {
return
}
stats := st.PoolStats()
r.opts.Meter.Counter(PoolHitsTotal).Set(uint64(stats.Hits))
r.opts.Meter.Counter(PoolMissesTotal).Set(uint64(stats.Misses))
r.opts.Meter.Counter(PoolTimeoutTotal).Set(uint64(stats.Timeouts))
r.opts.Meter.Counter(PoolConnTotalCurrent).Set(uint64(stats.TotalConns))
r.opts.Meter.Counter(PoolConnIdleCurrent).Set(uint64(stats.IdleConns))
r.opts.Meter.Counter(PoolConnStaleTotal).Set(uint64(stats.StaleConns))
} }
stats := st.PoolStats()
r.opts.Meter.Counter(PoolHitsTotal).Set(uint64(stats.Hits))
r.opts.Meter.Counter(PoolMissesTotal).Set(uint64(stats.Misses))
r.opts.Meter.Counter(PoolTimeoutTotal).Set(uint64(stats.Timeouts))
r.opts.Meter.Counter(PoolConnTotalCurrent).Set(uint64(stats.TotalConns))
r.opts.Meter.Counter(PoolConnIdleCurrent).Set(uint64(stats.IdleConns))
r.opts.Meter.Counter(PoolConnStaleTotal).Set(uint64(stats.StaleConns))
} }
}() }()
} }

View File

@ -6,27 +6,25 @@ import (
"net" "net"
"strconv" "strconv"
"github.com/redis/go-redis/extra/rediscmd/v9" rediscmd "github.com/redis/go-redis/extra/rediscmd/v9"
"github.com/redis/go-redis/v9" goredis "github.com/redis/go-redis/v9"
"go.unistack.org/micro/v3/tracer" "go.unistack.org/micro/v3/tracer"
) )
func setTracing(rdb redis.UniversalClient, tr tracer.Tracer, opts ...tracer.SpanOption) { func setTracing(rdb goredis.UniversalClient, tr tracer.Tracer, opts ...tracer.SpanOption) {
switch rdb := rdb.(type) { switch rdb := rdb.(type) {
case *redis.Client: case *goredis.Client:
opt := rdb.Options() opt := rdb.Options()
connString := formatDBConnString(opt.Network, opt.Addr) connString := formatDBConnString(opt.Network, opt.Addr)
rdb.AddHook(newTracingHook(connString, tr)) rdb.AddHook(newTracingHook(connString, tr))
case *redis.ClusterClient: case *goredis.ClusterClient:
rdb.AddHook(newTracingHook("", tr, opts...)) rdb.OnNewNode(func(rdb *goredis.Client) {
rdb.OnNewNode(func(rdb *redis.Client) {
opt := rdb.Options() opt := rdb.Options()
connString := formatDBConnString(opt.Network, opt.Addr) connString := formatDBConnString(opt.Network, opt.Addr)
rdb.AddHook(newTracingHook(connString, tr)) rdb.AddHook(newTracingHook(connString, tr))
}) })
case *redis.Ring: case *goredis.Ring:
rdb.AddHook(newTracingHook("", tr, opts...)) rdb.OnNewNode(func(rdb *goredis.Client) {
rdb.OnNewNode(func(rdb *redis.Client) {
opt := rdb.Options() opt := rdb.Options()
connString := formatDBConnString(opt.Network, opt.Addr) connString := formatDBConnString(opt.Network, opt.Addr)
rdb.AddHook(newTracingHook(connString, tr)) rdb.AddHook(newTracingHook(connString, tr))
@ -39,7 +37,7 @@ type tracingHook struct {
opts []tracer.SpanOption opts []tracer.SpanOption
} }
var _ redis.Hook = (*tracingHook)(nil) var _ goredis.Hook = (*tracingHook)(nil)
func newTracingHook(connString string, tr tracer.Tracer, opts ...tracer.SpanOption) *tracingHook { func newTracingHook(connString string, tr tracer.Tracer, opts ...tracer.SpanOption) *tracingHook {
opts = append(opts, tracer.WithSpanKind(tracer.SpanKindClient)) opts = append(opts, tracer.WithSpanKind(tracer.SpanKindClient))
@ -53,45 +51,54 @@ func newTracingHook(connString string, tr tracer.Tracer, opts ...tracer.SpanOpti
} }
} }
func (h *tracingHook) DialHook(hook redis.DialHook) redis.DialHook { func (h *tracingHook) DialHook(hook goredis.DialHook) goredis.DialHook {
return func(ctx context.Context, network, addr string) (net.Conn, error) { return func(ctx context.Context, network, addr string) (net.Conn, error) {
sctx, span := h.tr.Start(ctx, "redis.dial", h.opts...) /*
defer span.Finish() _, span := h.tr.Start(ctx, "goredis.dial", h.opts...)
defer span.Finish()
conn, err := hook(sctx, network, addr) */
recordError(span, err) conn, err := hook(ctx, network, addr)
// recordError(span, err)
return conn, err return conn, err
} }
} }
func (h *tracingHook) ProcessHook(hook redis.ProcessHook) redis.ProcessHook { func (h *tracingHook) ProcessHook(hook goredis.ProcessHook) goredis.ProcessHook {
return func(ctx context.Context, cmd redis.Cmder) error { return func(ctx context.Context, cmd goredis.Cmder) error {
cmdString := rediscmd.CmdString(cmd) cmdString := rediscmd.CmdString(cmd)
var err error
sctx, span := h.tr.Start(ctx, "redis.process", append(h.opts, tracer.WithSpanLabels("db.statement", cmdString))...) switch cmdString {
defer span.Finish() case "cluster slots":
break
default:
_, span := h.tr.Start(ctx, "goredis.process", append(h.opts, tracer.WithSpanLabels("db.statement", cmdString))...)
defer func() {
recordError(span, err)
span.Finish()
}()
}
err := hook(sctx, cmd) err = hook(ctx, cmd)
recordError(span, err)
return err return err
} }
} }
func (h *tracingHook) ProcessPipelineHook(hook redis.ProcessPipelineHook) redis.ProcessPipelineHook { func (h *tracingHook) ProcessPipelineHook(hook goredis.ProcessPipelineHook) goredis.ProcessPipelineHook {
return func(ctx context.Context, cmds []redis.Cmder) error { return func(ctx context.Context, cmds []goredis.Cmder) error {
_, cmdsString := rediscmd.CmdsString(cmds) _, cmdsString := rediscmd.CmdsString(cmds)
opts := append(h.opts, tracer.WithSpanLabels( opts := append(h.opts, tracer.WithSpanLabels(
"db.redis.num_cmd", strconv.Itoa(len(cmds)), "db.goredis.num_cmd", strconv.Itoa(len(cmds)),
"db.statement", cmdsString, "db.statement", cmdsString,
)) ))
sctx, span := h.tr.Start(ctx, "redis.process_pipeline", opts...) _, span := h.tr.Start(ctx, "goredis.process_pipeline", opts...)
defer span.Finish() defer span.Finish()
err := hook(sctx, cmds) err := hook(ctx, cmds)
recordError(span, err) recordError(span, err)
return err return err
@ -99,7 +106,7 @@ func (h *tracingHook) ProcessPipelineHook(hook redis.ProcessPipelineHook) redis.
} }
func setSpanError(ctx context.Context, err error) { func setSpanError(ctx context.Context, err error) {
if err == nil || err == redis.Nil { if err == nil || err == goredis.Nil {
return return
} }
if sp, ok := tracer.SpanFromContext(ctx); !ok && sp != nil { if sp, ok := tracer.SpanFromContext(ctx); !ok && sp != nil {
@ -108,7 +115,7 @@ func setSpanError(ctx context.Context, err error) {
} }
func recordError(span tracer.Span, err error) { func recordError(span tracer.Span, err error) {
if err != nil && err != redis.Nil { if err != nil && err != goredis.Nil {
span.SetStatus(tracer.SpanStatusError, err.Error()) span.SetStatus(tracer.SpanStatusError, err.Error())
} }
} }