Compare commits
	
		
			184 Commits
		
	
	
		
	
	| Author | SHA1 | Date | |
|---|---|---|---|
| 4bb73514e9 | |||
| 3e86864ce7 | |||
|  | a68d3b24b8 | ||
| 9c22ae5384 | |||
|  | 16bad9a0cd | ||
|  | 3c779b248f | ||
| bbc7512054 | |||
|  | dd810e4ae0 | ||
| 236ed47ab1 | |||
|  | 909bcf51a4 | ||
| 9c24001f52 | |||
| 680cd6f708 | |||
|  | 3fcf3bef6d | ||
|  | 0ecd6199d4 | ||
| 14a30fb6a7 | |||
| 7c613072df | |||
|  | c55e212270 | ||
| 100bc006bb | |||
| 5dbfe8a7a6 | |||
| 6be077dbe8 | |||
| b4878211ee | |||
| ec9178c6d4 | |||
| ae63d44866 | |||
| 883e79216a | |||
| fa636ef6a9 | |||
| cdb81a9ba3 | |||
| 413c6cc2f0 | |||
|  | f56bd70136 | ||
| b51b4107a8 | |||
| 2067c9de6b | |||
| 3f82cb3ba4 | |||
|  | 306b7a3962 | ||
| a8eda9d58d | |||
| 7e4477dcb4 | |||
|  | d846044fc6 | ||
| 29d956e74e | |||
| fcc4faff8a | |||
| 5df8f83f45 | |||
|  | 27fa6e9173 | ||
| bd55a35dc3 | |||
| 653bd386cc | |||
|  | 558c6f4d7c | ||
| d7dd6fbeb2 | |||
| a00cf2c8d9 | |||
|  | a3e8ab2492 | ||
| 06da500ef4 | |||
| 277f04ba19 | |||
|  | 470263ff5f | ||
| b8232e02be | |||
|  | f8c5e10c1d | ||
| 397e71f815 | |||
| 74e31d99f6 | |||
| f39de15d93 | |||
| d291102877 | |||
| 37ffbb18d8 | |||
| 9a85dead86 | |||
| a489aab1c3 | |||
| d160664ef1 | |||
| fa868edcaa | |||
|  | 6ed0b0e090 | ||
| 533b265d19 | |||
| 1ace2631a4 | |||
| 3dd5ca68d1 | |||
| 66ccd6021f | |||
| e5346f7e4f | |||
|  | daf19f031a | ||
| 5989fd54ca | |||
| ed30c26324 | |||
| 0f8f93d09a | |||
|  | f460e2f8dd | ||
| 70d6a79274 | |||
| 664b1586af | |||
| 8d747c64a8 | |||
| 94beb5ed3b | |||
| 98981ba86c | |||
| 1013f50d0e | |||
| 0b190997b1 | |||
| 69a44eb190 | |||
| 0476028f69 | |||
| 330d8b149a | |||
| 19b04fe070 | |||
| 4cd55875c6 | |||
| a7896cc728 | |||
| ff991bf49c | |||
| 5a6551b703 | |||
| 9406a33d60 | |||
| 8f185abd9d | |||
| 86492e0644 | |||
| b21972964a | |||
| f5ee065d09 | |||
| 8cb02f2b08 | |||
| bc926cd6bd | |||
| 356abfd818 | |||
| 18444d3f98 | |||
| d5f07922e8 | |||
| 675e717410 | |||
| 7b6aea235a | |||
| 2cb7200467 | |||
| f430f97a97 | |||
| 0060c4377a | |||
| e46579fe9a | |||
| ca52973194 | |||
| 5bb33c7e1d | |||
| b71fc25328 | |||
| 9345dd075a | |||
| 1c1b9c0a28 | |||
| 2969494c5a | |||
| cbd3fa38ba | |||
| 569a36383d | |||
| 90bed77526 | |||
| ba4478a5e0 | |||
| 6dc76cdfea | |||
| e266683d96 | |||
| 2b62ad04f2 | |||
| 275b0a64e5 | |||
| 38c5fe8b5a | |||
| b6a0e4d983 | |||
| d9b822deff | |||
| 0e66688f8f | |||
| 9213fd212f | |||
| aa360dcf51 | |||
| 2df259b5b8 | |||
| 15e9310368 | |||
|  | 16d8cf3434 | ||
| 9704ef2e5e | |||
| 94e8f90f00 | |||
| 34d1587881 | |||
| bf4143cde5 | |||
| 36b7b9f5fb | |||
| ae97023092 | |||
| 115ca6a018 | |||
| 89cf4ef8af | |||
| 2a6ce6d4da | |||
| ad19fe2b90 | |||
| 49055a28ea | |||
| d1c6e121c1 | |||
| 7cd7fb0c0a | |||
| 77eb5b5264 | |||
| 929e46c087 | |||
| 1fb5673d27 | |||
| 3bbb0cbc72 | |||
| 71fe0df73f | |||
| f1b8ecbdb3 | |||
| fd2b2762e9 | |||
| 82d269cfb4 | |||
| 6641463eed | |||
| faf2454f0a | |||
| de9e4d73f5 | |||
| 4ae7277140 | |||
| a98618ed5b | |||
| 3aaf1182cb | |||
| eb1482d789 | |||
| a305f7553f | |||
|  | d9b2f2a45d | ||
| 3ace7657dc | |||
| 53b40617e2 | |||
| 1a9236caad | |||
| 6c68d39081 | |||
| 35e62fbeb0 | |||
| 00b3ceb468 | |||
| 7dc8f088c9 | |||
| c65afcea1b | |||
| 3eebfb5b11 | |||
| fa1427014c | |||
| 62074965ee | |||
| 9c8fbb2202 | |||
| 7c0a5f5e2a | |||
| b08f5321b0 | |||
| cc0f24e012 | |||
| 307a08f50c | |||
| edc93e8c37 | |||
| 391813c260 | |||
| 1a1459dd0e | |||
| 4e99680c30 | |||
| 92a3a547b8 | |||
| 849c462037 | |||
| 54a55c83e2 | |||
| 781dee03db | |||
| 26dd2eb405 | |||
| 3a21069b86 | |||
| add3ce478c | |||
|  | c3de003e4a | ||
| 7b7cf18a65 | |||
| 1bcf71c189 | 
| @@ -1,24 +0,0 @@ | ||||
| name: lint | ||||
| on: | ||||
|   pull_request: | ||||
|     branches: | ||||
|     - master | ||||
|     - v3 | ||||
| jobs: | ||||
|   lint: | ||||
|     name: lint | ||||
|     runs-on: ubuntu-latest | ||||
|     steps: | ||||
|     - name: setup-go | ||||
|       uses: actions/setup-go@v3 | ||||
|       with: | ||||
|         go-version: 1.21 | ||||
|     - name: checkout | ||||
|       uses: actions/checkout@v3 | ||||
|     - name: deps | ||||
|       run: go get -v -d ./... | ||||
|     - name: lint | ||||
|       uses: https://github.com/golangci/golangci-lint-action@v3.4.0 | ||||
|       continue-on-error: true | ||||
|       with: | ||||
|         version: v1.52 | ||||
| @@ -1,23 +0,0 @@ | ||||
| name: pr | ||||
| on: | ||||
|   pull_request: | ||||
|     branches: | ||||
|     - master | ||||
|     - v3 | ||||
| jobs: | ||||
|   test: | ||||
|     name: test | ||||
|     runs-on: ubuntu-latest | ||||
|     steps: | ||||
|     - name: checkout | ||||
|       uses: actions/checkout@v3 | ||||
|     - name: setup-go | ||||
|       uses: actions/setup-go@v3 | ||||
|       with: | ||||
|         go-version: 1.21 | ||||
|     - name: deps | ||||
|       run: go get -v -t -d ./... | ||||
|     - name: test | ||||
|       env: | ||||
|         INTEGRATION_TESTS: yes | ||||
|       run: go test -mod readonly -v ./... | ||||
							
								
								
									
										53
									
								
								.github/workflows/job_coverage.yml
									
									
									
									
										vendored
									
									
										Normal file
									
								
							
							
						
						
									
										53
									
								
								.github/workflows/job_coverage.yml
									
									
									
									
										vendored
									
									
										Normal file
									
								
							| @@ -0,0 +1,53 @@ | ||||
| name: coverage | ||||
|  | ||||
| on: | ||||
|   push: | ||||
|     branches: [ main, v3, v4 ] | ||||
|     paths-ignore: | ||||
|       - '.github/**' | ||||
|       - '.gitea/**' | ||||
|   pull_request: | ||||
|     branches: [ main, v3, v4 ] | ||||
|  | ||||
| jobs: | ||||
|  | ||||
|   build: | ||||
|     if: github.server_url != 'https://github.com' | ||||
|     runs-on: ubuntu-latest | ||||
|     steps: | ||||
|     - name: checkout code | ||||
|       uses: actions/checkout@v4 | ||||
|       with: | ||||
|         filter: 'blob:none' | ||||
|  | ||||
|     - name: setup go | ||||
|       uses: actions/setup-go@v5 | ||||
|       with: | ||||
|         cache-dependency-path: "**/*.sum" | ||||
|         go-version: 'stable' | ||||
|  | ||||
|     - name: test coverage | ||||
|       run: | | ||||
|         go test -v -cover ./... -covermode=count -coverprofile coverage.out -coverpkg ./... | ||||
|         go tool cover -func coverage.out -o coverage.out | ||||
|  | ||||
|     - name: coverage badge | ||||
|       uses: tj-actions/coverage-badge-go@v2 | ||||
|       with: | ||||
|         green: 80 | ||||
|         filename: coverage.out | ||||
|  | ||||
|     - uses: stefanzweifel/git-auto-commit-action@v4 | ||||
|       name: autocommit | ||||
|       with: | ||||
|         commit_message: Apply Code Coverage Badge | ||||
|         skip_fetch: false | ||||
|         skip_checkout: false | ||||
|         file_pattern: ./README.md | ||||
|  | ||||
|     - name: push | ||||
|       if: steps.auto-commit-action.outputs.changes_detected == 'true' | ||||
|       uses: ad-m/github-push-action@master | ||||
|       with: | ||||
|         github_token: ${{ github.token }} | ||||
|         branch: ${{ github.ref }} | ||||
							
								
								
									
										29
									
								
								.github/workflows/job_lint.yml
									
									
									
									
										vendored
									
									
										Normal file
									
								
							
							
						
						
									
										29
									
								
								.github/workflows/job_lint.yml
									
									
									
									
										vendored
									
									
										Normal file
									
								
							| @@ -0,0 +1,29 @@ | ||||
| name: lint | ||||
|  | ||||
| on: | ||||
|   pull_request: | ||||
|     types: [opened, reopened, synchronize] | ||||
|     branches: [ master, v3, v4 ] | ||||
|     paths-ignore: | ||||
|       - '.github/**' | ||||
|       - '.gitea/**' | ||||
|  | ||||
| jobs: | ||||
|   lint: | ||||
|     runs-on: ubuntu-latest | ||||
|     steps: | ||||
|     - name: checkout code | ||||
|       uses: actions/checkout@v4 | ||||
|       with: | ||||
|         filter: 'blob:none' | ||||
|     - name: setup go | ||||
|       uses: actions/setup-go@v5 | ||||
|       with: | ||||
|         cache-dependency-path: "**/*.sum" | ||||
|         go-version: 'stable' | ||||
|     - name: setup deps | ||||
|       run: go get -v ./... | ||||
|     - name: run lint | ||||
|       uses: golangci/golangci-lint-action@v6 | ||||
|       with: | ||||
|         version: 'latest' | ||||
							
								
								
									
										31
									
								
								.github/workflows/job_test.yml
									
									
									
									
										vendored
									
									
										Normal file
									
								
							
							
						
						
									
										31
									
								
								.github/workflows/job_test.yml
									
									
									
									
										vendored
									
									
										Normal file
									
								
							| @@ -0,0 +1,31 @@ | ||||
| name: test | ||||
|  | ||||
| on: | ||||
|   pull_request: | ||||
|     types: [opened, reopened, synchronize] | ||||
|     branches: [ master, v3, v4 ] | ||||
|   push: | ||||
|     branches: [ master, v3, v4 ] | ||||
|     paths-ignore: | ||||
|       - '.github/**' | ||||
|       - '.gitea/**' | ||||
|  | ||||
| jobs: | ||||
|   test: | ||||
|     runs-on: ubuntu-latest | ||||
|     steps: | ||||
|     - name: checkout code | ||||
|       uses: actions/checkout@v4 | ||||
|       with: | ||||
|         filter: 'blob:none' | ||||
|     - name: setup go | ||||
|       uses: actions/setup-go@v5 | ||||
|       with: | ||||
|         cache-dependency-path: "**/*.sum" | ||||
|         go-version: 'stable' | ||||
|     - name: setup deps | ||||
|       run: go get -v ./... | ||||
|     - name: run test | ||||
|       env: | ||||
|         INTEGRATION_TESTS: yes | ||||
|       run: go test -mod readonly -v ./... | ||||
							
								
								
									
										50
									
								
								.github/workflows/job_tests.yml
									
									
									
									
										vendored
									
									
										Normal file
									
								
							
							
						
						
									
										50
									
								
								.github/workflows/job_tests.yml
									
									
									
									
										vendored
									
									
										Normal file
									
								
							| @@ -0,0 +1,50 @@ | ||||
| name: test | ||||
|  | ||||
| on: | ||||
|   pull_request: | ||||
|     types: [opened, reopened, synchronize] | ||||
|     branches: [ master, v3, v4 ] | ||||
|   push: | ||||
|     branches: [ master, v3, v4 ] | ||||
|     paths-ignore: | ||||
|       - '.github/**' | ||||
|       - '.gitea/**' | ||||
|  | ||||
| jobs: | ||||
|   test: | ||||
|     runs-on: ubuntu-latest | ||||
|     steps: | ||||
|     - name: checkout code | ||||
|       uses: actions/checkout@v4 | ||||
|       with: | ||||
|         filter: 'blob:none' | ||||
|     - name: checkout tests | ||||
|       uses: actions/checkout@v4 | ||||
|       with: | ||||
|         ref: master | ||||
|         filter: 'blob:none' | ||||
|         repository: unistack-org/micro-tests | ||||
|         path: micro-tests | ||||
|     - name: setup go | ||||
|       uses: actions/setup-go@v5 | ||||
|       with: | ||||
|         cache-dependency-path: "**/*.sum" | ||||
|         go-version: 'stable' | ||||
|     - name: setup go work | ||||
|       env: | ||||
|         GOWORK: ${{ github.workspace }}/go.work | ||||
|       run: | | ||||
|         go work init | ||||
|         go work use . | ||||
|         go work use micro-tests | ||||
|     - name: setup deps | ||||
|       env: | ||||
|         GOWORK: ${{ github.workspace }}/go.work | ||||
|       run: go get -v ./... | ||||
|     - name: run tests | ||||
|       env: | ||||
|         INTEGRATION_TESTS: yes | ||||
|         GOWORK: ${{ github.workspace }}/go.work | ||||
|       run: | | ||||
|         cd micro-tests | ||||
|         go test -mod readonly -v ./... || true | ||||
| @@ -1,44 +1,5 @@ | ||||
| run: | ||||
|   concurrency: 4 | ||||
|   deadline: 5m | ||||
|   concurrency: 8 | ||||
|   timeout: 5m | ||||
|   issues-exit-code: 1 | ||||
|   tests: true | ||||
|  | ||||
| linters-settings: | ||||
|   govet: | ||||
|     check-shadowing: true | ||||
|     enable: | ||||
|       - fieldalignment | ||||
|  | ||||
| linters: | ||||
|   enable: | ||||
|     - govet | ||||
|     - deadcode | ||||
|     - errcheck | ||||
|     - govet | ||||
|     - ineffassign | ||||
|     - staticcheck | ||||
|     - structcheck | ||||
|     - typecheck | ||||
|     - unused | ||||
|     - varcheck | ||||
|     - bodyclose | ||||
|     - gci | ||||
|     - goconst | ||||
|     - gocritic | ||||
|     - gosimple | ||||
|     - gofmt | ||||
|     - gofumpt | ||||
|     - goimports | ||||
|     - revive | ||||
|     - gosec | ||||
|     - makezero | ||||
|     - misspell | ||||
|     - nakedret | ||||
|     - nestif | ||||
|     - nilerr | ||||
|     - noctx | ||||
|     - prealloc | ||||
|     - unconvert | ||||
|     - unparam | ||||
|   disable-all: false | ||||
|   | ||||
							
								
								
									
										33
									
								
								README.md
									
									
									
									
									
								
							
							
						
						
									
										33
									
								
								README.md
									
									
									
									
									
								
							| @@ -1,39 +1,34 @@ | ||||
| # Micro [](https://opensource.org/licenses/Apache-2.0) [](https://pkg.go.dev/github.com/unistack-org/micro/v3?tab=overview) [](https://github.com/unistack-org/micro/actions?query=workflow%3Abuild+branch%3Amaster+event%3Apush) [](https://goreportcard.com/report/go.unistack.org/micro/v3) [](https://codecov.io/gh/unistack-org/micro) | ||||
| # Micro | ||||
|  | ||||
| [](https://opensource.org/licenses/Apache-2.0) | ||||
| [](https://pkg.go.dev/go.unistack.org/micro/v3?tab=overview) | ||||
| [](https://git.unistack.org/unistack-org/micro/actions?query=workflow%3Abuild+branch%3Av3+event%3Apush) | ||||
| [](https://goreportcard.com/report/go.unistack.org/micro/v3) | ||||
|  | ||||
| Micro is a standard library for microservices. | ||||
|  | ||||
| ## Overview | ||||
|  | ||||
| Micro provides the core requirements for distributed systems development including RPC and Event driven communication.  | ||||
| Micro provides the core requirements for distributed systems development including SYNC and ASYNC communication.  | ||||
|  | ||||
| ## Features | ||||
|  | ||||
| Micro abstracts away the details of distributed systems. Here are the main features. | ||||
|  | ||||
| - **Authentication** - Auth is built in as a first class citizen. Authentication and authorization enable secure  | ||||
| zero trust networking by providing every service an identity and certificates. This additionally includes rule  | ||||
| based access control. | ||||
| Micro abstracts away the details of distributed systems. Main features: | ||||
|  | ||||
| - **Dynamic Config** - Load and hot reload dynamic config from anywhere. The config interface provides a way to load application  | ||||
| level config from any source such as env vars, file, etcd. You can merge the sources and even define fallbacks. | ||||
| level config from any source such as env vars, cmdline, file, consul, vault, etc... You can merge the sources and even define fallbacks. | ||||
|  | ||||
| - **Data Storage** - A simple data store interface to read, write and delete records. It includes support for memory, file and  | ||||
| CockroachDB by default. State and persistence becomes a core requirement beyond prototyping and Micro looks to build that into the framework. | ||||
| s3. State and persistence becomes a core requirement beyond prototyping and Micro looks to build that into the framework. | ||||
|  | ||||
| - **Service Discovery** - Automatic service registration and name resolution. Service discovery is at the core of micro service  | ||||
| development. When service A needs to speak to service B it needs the location of that service. | ||||
|  | ||||
| - **Load Balancing** - Client side load balancing built on service discovery. Once we have the addresses of any number of instances  | ||||
| of a service we now need a way to decide which node to route to. We use random hashed load balancing to provide even distribution  | ||||
| across the services and retry a different node if there's a problem.  | ||||
| development. | ||||
|  | ||||
| - **Message Encoding** - Dynamic message encoding based on content-type. The client and server will use codecs along with content-type  | ||||
| to seamlessly encode and decode Go types for you. Any variety of messages could be encoded and sent from different clients. The client  | ||||
| and server handle this by default. | ||||
|  | ||||
| - **Transport** - gRPC or http based request/response with support for bidirectional streaming. We provide an abstraction for synchronous communication. A request made to a service will be automatically resolved, load balanced, dialled and streamed. | ||||
|  | ||||
| - **Async Messaging** - PubSub is built in as a first class citizen for asynchronous communication and event driven architectures. | ||||
| - **Async Messaging** - Pub/Sub is built in as a first class citizen for asynchronous communication and event driven architectures. | ||||
| Event notifications are a core pattern in micro service development. | ||||
|  | ||||
| - **Synchronization** - Distributed systems are often built in an eventually consistent manner. Support for distributed locking and  | ||||
| @@ -42,10 +37,6 @@ leadership are built in as a Sync interface. When using an eventually consistent | ||||
| - **Pluggable Interfaces** - Micro makes use of Go interfaces for each system abstraction. Because of this these interfaces  | ||||
| are pluggable and allows Micro to be runtime agnostic. | ||||
|  | ||||
| ## Getting Started | ||||
|  | ||||
| To be created. | ||||
|  | ||||
| ## License | ||||
|  | ||||
| Micro is Apache 2.0 licensed. | ||||
|   | ||||
							
								
								
									
										15
									
								
								SECURITY.md
									
									
									
									
									
								
							
							
						
						
									
										15
									
								
								SECURITY.md
									
									
									
									
									
								
							| @@ -1,15 +0,0 @@ | ||||
| # Security Policy | ||||
|  | ||||
| ## Supported Versions | ||||
|  | ||||
| Use this section to tell people about which versions of your project are | ||||
| currently being supported with security updates. | ||||
|  | ||||
| | Version | Supported          | | ||||
| | ------- | ------------------ | | ||||
| | 3.7.x   | :white_check_mark: | | ||||
| | < 3.7.0 | :x:                | | ||||
|  | ||||
| ## Reporting a Vulnerability | ||||
|  | ||||
| If you find any issue, please create github issue in this repo | ||||
| @@ -1,5 +1,5 @@ | ||||
| // Package broker is an interface used for asynchronous messaging | ||||
| package broker // import "go.unistack.org/micro/v3/broker" | ||||
| package broker | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| @@ -46,8 +46,25 @@ type Broker interface { | ||||
| 	BatchSubscribe(ctx context.Context, topic string, h BatchHandler, opts ...SubscribeOption) (Subscriber, error) | ||||
| 	// String type of broker | ||||
| 	String() string | ||||
| 	// Live returns broker liveness | ||||
| 	Live() bool | ||||
| 	// Ready returns broker readiness | ||||
| 	Ready() bool | ||||
| 	// Health returns broker health | ||||
| 	Health() bool | ||||
| } | ||||
|  | ||||
| type ( | ||||
| 	FuncPublish        func(ctx context.Context, topic string, msg *Message, opts ...PublishOption) error | ||||
| 	HookPublish        func(next FuncPublish) FuncPublish | ||||
| 	FuncBatchPublish   func(ctx context.Context, msgs []*Message, opts ...PublishOption) error | ||||
| 	HookBatchPublish   func(next FuncBatchPublish) FuncBatchPublish | ||||
| 	FuncSubscribe      func(ctx context.Context, topic string, h Handler, opts ...SubscribeOption) (Subscriber, error) | ||||
| 	HookSubscribe      func(next FuncSubscribe) FuncSubscribe | ||||
| 	FuncBatchSubscribe func(ctx context.Context, topic string, h BatchHandler, opts ...SubscribeOption) (Subscriber, error) | ||||
| 	HookBatchSubscribe func(next FuncBatchSubscribe) FuncBatchSubscribe | ||||
| ) | ||||
|  | ||||
| // Handler is used to process messages via a subscription of a topic. | ||||
| type Handler func(Event) error | ||||
|  | ||||
| @@ -77,6 +94,8 @@ type BatchHandler func(Events) error | ||||
|  | ||||
| // Event is given to a subscription handler for processing | ||||
| type Event interface { | ||||
| 	// Context return context.Context for event | ||||
| 	Context() context.Context | ||||
| 	// Topic returns event topic | ||||
| 	Topic() string | ||||
| 	// Message returns broker message | ||||
|   | ||||
| @@ -15,6 +15,15 @@ func FromContext(ctx context.Context) (Broker, bool) { | ||||
| 	return c, ok | ||||
| } | ||||
|  | ||||
| // MustContext returns broker from passed context | ||||
| func MustContext(ctx context.Context) Broker { | ||||
| 	b, ok := FromContext(ctx) | ||||
| 	if !ok { | ||||
| 		panic("missing broker") | ||||
| 	} | ||||
| 	return b | ||||
| } | ||||
|  | ||||
| // NewContext savess broker in context | ||||
| func NewContext(ctx context.Context, s Broker) context.Context { | ||||
| 	if ctx == nil { | ||||
|   | ||||
| @@ -7,6 +7,7 @@ import ( | ||||
| 	"go.unistack.org/micro/v3/broker" | ||||
| 	"go.unistack.org/micro/v3/logger" | ||||
| 	"go.unistack.org/micro/v3/metadata" | ||||
| 	"go.unistack.org/micro/v3/options" | ||||
| 	maddr "go.unistack.org/micro/v3/util/addr" | ||||
| 	"go.unistack.org/micro/v3/util/id" | ||||
| 	mnet "go.unistack.org/micro/v3/util/net" | ||||
| @@ -14,9 +15,13 @@ import ( | ||||
| ) | ||||
|  | ||||
| type memoryBroker struct { | ||||
| 	subscribers map[string][]*memorySubscriber | ||||
| 	addr        string | ||||
| 	opts        broker.Options | ||||
| 	funcPublish        broker.FuncPublish | ||||
| 	funcBatchPublish   broker.FuncBatchPublish | ||||
| 	funcSubscribe      broker.FuncSubscribe | ||||
| 	funcBatchSubscribe broker.FuncBatchSubscribe | ||||
| 	subscribers        map[string][]*memorySubscriber | ||||
| 	addr               string | ||||
| 	opts               broker.Options | ||||
| 	sync.RWMutex | ||||
| 	connected bool | ||||
| } | ||||
| @@ -98,15 +103,42 @@ func (m *memoryBroker) Init(opts ...broker.Option) error { | ||||
| 	for _, o := range opts { | ||||
| 		o(&m.opts) | ||||
| 	} | ||||
|  | ||||
| 	m.funcPublish = m.fnPublish | ||||
| 	m.funcBatchPublish = m.fnBatchPublish | ||||
| 	m.funcSubscribe = m.fnSubscribe | ||||
| 	m.funcBatchSubscribe = m.fnBatchSubscribe | ||||
|  | ||||
| 	m.opts.Hooks.EachPrev(func(hook options.Hook) { | ||||
| 		switch h := hook.(type) { | ||||
| 		case broker.HookPublish: | ||||
| 			m.funcPublish = h(m.funcPublish) | ||||
| 		case broker.HookBatchPublish: | ||||
| 			m.funcBatchPublish = h(m.funcBatchPublish) | ||||
| 		case broker.HookSubscribe: | ||||
| 			m.funcSubscribe = h(m.funcSubscribe) | ||||
| 		case broker.HookBatchSubscribe: | ||||
| 			m.funcBatchSubscribe = h(m.funcBatchSubscribe) | ||||
| 		} | ||||
| 	}) | ||||
|  | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| func (m *memoryBroker) Publish(ctx context.Context, topic string, msg *broker.Message, opts ...broker.PublishOption) error { | ||||
| 	return m.funcPublish(ctx, topic, msg, opts...) | ||||
| } | ||||
|  | ||||
| func (m *memoryBroker) fnPublish(ctx context.Context, topic string, msg *broker.Message, opts ...broker.PublishOption) error { | ||||
| 	msg.Header.Set(metadata.HeaderTopic, topic) | ||||
| 	return m.publish(ctx, []*broker.Message{msg}, opts...) | ||||
| } | ||||
|  | ||||
| func (m *memoryBroker) BatchPublish(ctx context.Context, msgs []*broker.Message, opts ...broker.PublishOption) error { | ||||
| 	return m.funcBatchPublish(ctx, msgs, opts...) | ||||
| } | ||||
|  | ||||
| func (m *memoryBroker) fnBatchPublish(ctx context.Context, msgs []*broker.Message, opts ...broker.PublishOption) error { | ||||
| 	return m.publish(ctx, msgs, opts...) | ||||
| } | ||||
|  | ||||
| @@ -174,7 +206,7 @@ func (m *memoryBroker) publish(ctx context.Context, msgs []*broker.Message, opts | ||||
| 						} | ||||
| 					} else if sub.opts.AutoAck { | ||||
| 						if err = ms.Ack(); err != nil { | ||||
| 							m.opts.Logger.Errorf(m.opts.Context, "ack failed: %v", err) | ||||
| 							m.opts.Logger.Error(m.opts.Context, "broker ack error", err) | ||||
| 						} | ||||
| 					} | ||||
| 					// single processing | ||||
| @@ -185,11 +217,11 @@ func (m *memoryBroker) publish(ctx context.Context, msgs []*broker.Message, opts | ||||
| 							if eh != nil { | ||||
| 								_ = eh(p) | ||||
| 							} else if m.opts.Logger.V(logger.ErrorLevel) { | ||||
| 								m.opts.Logger.Error(m.opts.Context, err.Error()) | ||||
| 								m.opts.Logger.Error(m.opts.Context, "broker handler error", err) | ||||
| 							} | ||||
| 						} else if sub.opts.AutoAck { | ||||
| 							if err = p.Ack(); err != nil { | ||||
| 								m.opts.Logger.Errorf(m.opts.Context, "ack failed: %v", err) | ||||
| 								m.opts.Logger.Error(m.opts.Context, "broker ack error", err) | ||||
| 							} | ||||
| 						} | ||||
| 					} | ||||
| @@ -202,6 +234,10 @@ func (m *memoryBroker) publish(ctx context.Context, msgs []*broker.Message, opts | ||||
| } | ||||
|  | ||||
| func (m *memoryBroker) BatchSubscribe(ctx context.Context, topic string, handler broker.BatchHandler, opts ...broker.SubscribeOption) (broker.Subscriber, error) { | ||||
| 	return m.funcBatchSubscribe(ctx, topic, handler, opts...) | ||||
| } | ||||
|  | ||||
| func (m *memoryBroker) fnBatchSubscribe(ctx context.Context, topic string, handler broker.BatchHandler, opts ...broker.SubscribeOption) (broker.Subscriber, error) { | ||||
| 	m.RLock() | ||||
| 	if !m.connected { | ||||
| 		m.RUnlock() | ||||
| @@ -247,6 +283,10 @@ func (m *memoryBroker) BatchSubscribe(ctx context.Context, topic string, handler | ||||
| } | ||||
|  | ||||
| func (m *memoryBroker) Subscribe(ctx context.Context, topic string, handler broker.Handler, opts ...broker.SubscribeOption) (broker.Subscriber, error) { | ||||
| 	return m.funcSubscribe(ctx, topic, handler, opts...) | ||||
| } | ||||
|  | ||||
| func (m *memoryBroker) fnSubscribe(ctx context.Context, topic string, handler broker.Handler, opts ...broker.SubscribeOption) (broker.Subscriber, error) { | ||||
| 	m.RLock() | ||||
| 	if !m.connected { | ||||
| 		m.RUnlock() | ||||
| @@ -299,6 +339,18 @@ func (m *memoryBroker) Name() string { | ||||
| 	return m.opts.Name | ||||
| } | ||||
|  | ||||
| func (m *memoryBroker) Live() bool { | ||||
| 	return true | ||||
| } | ||||
|  | ||||
| func (m *memoryBroker) Ready() bool { | ||||
| 	return true | ||||
| } | ||||
|  | ||||
| func (m *memoryBroker) Health() bool { | ||||
| 	return true | ||||
| } | ||||
|  | ||||
| func (m *memoryEvent) Topic() string { | ||||
| 	return m.topic | ||||
| } | ||||
| @@ -333,6 +385,10 @@ func (m *memoryEvent) SetError(err error) { | ||||
| 	m.err = err | ||||
| } | ||||
|  | ||||
| func (m *memoryEvent) Context() context.Context { | ||||
| 	return m.opts.Context | ||||
| } | ||||
|  | ||||
| func (m *memorySubscriber) Options() broker.SubscribeOptions { | ||||
| 	return m.opts | ||||
| } | ||||
|   | ||||
| @@ -13,6 +13,10 @@ func TestMemoryBatchBroker(t *testing.T) { | ||||
| 	b := NewBroker() | ||||
| 	ctx := context.Background() | ||||
|  | ||||
| 	if err := b.Init(); err != nil { | ||||
| 		t.Fatalf("Unexpected init error %v", err) | ||||
| 	} | ||||
|  | ||||
| 	if err := b.Connect(ctx); err != nil { | ||||
| 		t.Fatalf("Unexpected connect error %v", err) | ||||
| 	} | ||||
| @@ -59,6 +63,10 @@ func TestMemoryBroker(t *testing.T) { | ||||
| 	b := NewBroker() | ||||
| 	ctx := context.Background() | ||||
|  | ||||
| 	if err := b.Init(); err != nil { | ||||
| 		t.Fatalf("Unexpected init error %v", err) | ||||
| 	} | ||||
|  | ||||
| 	if err := b.Connect(ctx); err != nil { | ||||
| 		t.Fatalf("Unexpected connect error %v", err) | ||||
| 	} | ||||
| @@ -66,7 +74,7 @@ func TestMemoryBroker(t *testing.T) { | ||||
| 	topic := "test" | ||||
| 	count := 10 | ||||
|  | ||||
| 	fn := func(p broker.Event) error { | ||||
| 	fn := func(_ broker.Event) error { | ||||
| 		return nil | ||||
| 	} | ||||
|  | ||||
|   | ||||
| @@ -3,17 +3,40 @@ package broker | ||||
| import ( | ||||
| 	"context" | ||||
| 	"strings" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/options" | ||||
| ) | ||||
|  | ||||
| type NoopBroker struct { | ||||
| 	opts Options | ||||
| 	funcPublish        FuncPublish | ||||
| 	funcBatchPublish   FuncBatchPublish | ||||
| 	funcSubscribe      FuncSubscribe | ||||
| 	funcBatchSubscribe FuncBatchSubscribe | ||||
| 	opts               Options | ||||
| } | ||||
|  | ||||
| func NewBroker(opts ...Option) *NoopBroker { | ||||
| 	b := &NoopBroker{opts: NewOptions(opts...)} | ||||
| 	b.funcPublish = b.fnPublish | ||||
| 	b.funcBatchPublish = b.fnBatchPublish | ||||
| 	b.funcSubscribe = b.fnSubscribe | ||||
| 	b.funcBatchSubscribe = b.fnBatchSubscribe | ||||
|  | ||||
| 	return b | ||||
| } | ||||
|  | ||||
| func (b *NoopBroker) Health() bool { | ||||
| 	return true | ||||
| } | ||||
|  | ||||
| func (b *NoopBroker) Live() bool { | ||||
| 	return true | ||||
| } | ||||
|  | ||||
| func (b *NoopBroker) Ready() bool { | ||||
| 	return true | ||||
| } | ||||
|  | ||||
| func (b *NoopBroker) Name() string { | ||||
| 	return b.opts.Name | ||||
| } | ||||
| @@ -30,6 +53,25 @@ func (b *NoopBroker) Init(opts ...Option) error { | ||||
| 	for _, opt := range opts { | ||||
| 		opt(&b.opts) | ||||
| 	} | ||||
|  | ||||
| 	b.funcPublish = b.fnPublish | ||||
| 	b.funcBatchPublish = b.fnBatchPublish | ||||
| 	b.funcSubscribe = b.fnSubscribe | ||||
| 	b.funcBatchSubscribe = b.fnBatchSubscribe | ||||
|  | ||||
| 	b.opts.Hooks.EachPrev(func(hook options.Hook) { | ||||
| 		switch h := hook.(type) { | ||||
| 		case HookPublish: | ||||
| 			b.funcPublish = h(b.funcPublish) | ||||
| 		case HookBatchPublish: | ||||
| 			b.funcBatchPublish = h(b.funcBatchPublish) | ||||
| 		case HookSubscribe: | ||||
| 			b.funcSubscribe = h(b.funcSubscribe) | ||||
| 		case HookBatchSubscribe: | ||||
| 			b.funcBatchSubscribe = h(b.funcBatchSubscribe) | ||||
| 		} | ||||
| 	}) | ||||
|  | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| @@ -45,14 +87,22 @@ func (b *NoopBroker) Address() string { | ||||
| 	return strings.Join(b.opts.Addrs, ",") | ||||
| } | ||||
|  | ||||
| func (b *NoopBroker) BatchPublish(_ context.Context, _ []*Message, _ ...PublishOption) error { | ||||
| func (b *NoopBroker) fnBatchPublish(_ context.Context, _ []*Message, _ ...PublishOption) error { | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| func (b *NoopBroker) Publish(_ context.Context, _ string, _ *Message, _ ...PublishOption) error { | ||||
| func (b *NoopBroker) BatchPublish(ctx context.Context, msgs []*Message, opts ...PublishOption) error { | ||||
| 	return b.funcBatchPublish(ctx, msgs, opts...) | ||||
| } | ||||
|  | ||||
| func (b *NoopBroker) fnPublish(_ context.Context, _ string, _ *Message, _ ...PublishOption) error { | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| func (b *NoopBroker) Publish(ctx context.Context, topic string, msg *Message, opts ...PublishOption) error { | ||||
| 	return b.funcPublish(ctx, topic, msg, opts...) | ||||
| } | ||||
|  | ||||
| type NoopSubscriber struct { | ||||
| 	ctx          context.Context | ||||
| 	topic        string | ||||
| @@ -61,14 +111,22 @@ type NoopSubscriber struct { | ||||
| 	opts         SubscribeOptions | ||||
| } | ||||
|  | ||||
| func (b *NoopBroker) BatchSubscribe(ctx context.Context, topic string, handler BatchHandler, opts ...SubscribeOption) (Subscriber, error) { | ||||
| func (b *NoopBroker) fnBatchSubscribe(ctx context.Context, topic string, handler BatchHandler, opts ...SubscribeOption) (Subscriber, error) { | ||||
| 	return &NoopSubscriber{ctx: ctx, topic: topic, opts: NewSubscribeOptions(opts...), batchHandler: handler}, nil | ||||
| } | ||||
|  | ||||
| func (b *NoopBroker) Subscribe(ctx context.Context, topic string, handler Handler, opts ...SubscribeOption) (Subscriber, error) { | ||||
| func (b *NoopBroker) BatchSubscribe(ctx context.Context, topic string, handler BatchHandler, opts ...SubscribeOption) (Subscriber, error) { | ||||
| 	return b.funcBatchSubscribe(ctx, topic, handler, opts...) | ||||
| } | ||||
|  | ||||
| func (b *NoopBroker) fnSubscribe(ctx context.Context, topic string, handler Handler, opts ...SubscribeOption) (Subscriber, error) { | ||||
| 	return &NoopSubscriber{ctx: ctx, topic: topic, opts: NewSubscribeOptions(opts...), handler: handler}, nil | ||||
| } | ||||
|  | ||||
| func (b *NoopBroker) Subscribe(ctx context.Context, topic string, handler Handler, opts ...SubscribeOption) (Subscriber, error) { | ||||
| 	return b.funcSubscribe(ctx, topic, handler, opts...) | ||||
| } | ||||
|  | ||||
| func (s *NoopSubscriber) Options() SubscribeOptions { | ||||
| 	return s.opts | ||||
| } | ||||
| @@ -77,6 +135,6 @@ func (s *NoopSubscriber) Topic() string { | ||||
| 	return s.topic | ||||
| } | ||||
|  | ||||
| func (s *NoopSubscriber) Unsubscribe(ctx context.Context) error { | ||||
| func (s *NoopSubscriber) Unsubscribe(_ context.Context) error { | ||||
| 	return nil | ||||
| } | ||||
|   | ||||
							
								
								
									
										35
									
								
								broker/noop_test.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										35
									
								
								broker/noop_test.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,35 @@ | ||||
| package broker | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"testing" | ||||
| ) | ||||
|  | ||||
| type testHook struct { | ||||
| 	f bool | ||||
| } | ||||
|  | ||||
| func (t *testHook) Publish1(fn FuncPublish) FuncPublish { | ||||
| 	return func(ctx context.Context, topic string, msg *Message, opts ...PublishOption) error { | ||||
| 		t.f = true | ||||
| 		return fn(ctx, topic, msg, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestNoopHook(t *testing.T) { | ||||
| 	h := &testHook{} | ||||
|  | ||||
| 	b := NewBroker(Hooks(HookPublish(h.Publish1))) | ||||
|  | ||||
| 	if err := b.Init(); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	if err := b.Publish(context.TODO(), "", nil); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	if !h.f { | ||||
| 		t.Fatal("hook not works") | ||||
| 	} | ||||
| } | ||||
| @@ -8,6 +8,7 @@ import ( | ||||
| 	"go.unistack.org/micro/v3/codec" | ||||
| 	"go.unistack.org/micro/v3/logger" | ||||
| 	"go.unistack.org/micro/v3/meter" | ||||
| 	"go.unistack.org/micro/v3/options" | ||||
| 	"go.unistack.org/micro/v3/register" | ||||
| 	"go.unistack.org/micro/v3/sync" | ||||
| 	"go.unistack.org/micro/v3/tracer" | ||||
| @@ -15,6 +16,9 @@ import ( | ||||
|  | ||||
| // Options struct | ||||
| type Options struct { | ||||
| 	// Name holds the broker name | ||||
| 	Name string | ||||
|  | ||||
| 	// Tracer used for tracing | ||||
| 	Tracer tracer.Tracer | ||||
| 	// Register can be used for clustering | ||||
| @@ -27,19 +31,24 @@ type Options struct { | ||||
| 	Meter meter.Meter | ||||
| 	// Context holds external options | ||||
| 	Context context.Context | ||||
|  | ||||
| 	// Wait waits for a collection of goroutines to finish | ||||
| 	Wait *sync.WaitGroup | ||||
| 	// TLSConfig holds tls.TLSConfig options | ||||
| 	TLSConfig *tls.Config | ||||
|  | ||||
| 	// ErrorHandler used when broker can't unmarshal incoming message | ||||
| 	ErrorHandler Handler | ||||
| 	// BatchErrorHandler used when broker can't unmashal incoming messages | ||||
| 	BatchErrorHandler BatchHandler | ||||
| 	// Name holds the broker name | ||||
| 	Name string | ||||
|  | ||||
| 	// Addrs holds the broker address | ||||
| 	Addrs []string | ||||
| 	// Hooks can be run before broker Publish/BatchPublish and | ||||
| 	// Subscribe/BatchSubscribe methods | ||||
| 	Hooks options.Hooks | ||||
|  | ||||
| 	Wait *sync.WaitGroup | ||||
|  | ||||
| 	// GracefulTimeout contains time to wait to finish in flight requests | ||||
| 	GracefulTimeout time.Duration | ||||
| } | ||||
|  | ||||
| @@ -230,6 +239,13 @@ func Name(n string) Option { | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Hooks sets hook runs before action | ||||
| func Hooks(h ...options.Hook) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Hooks = append(o.Hooks, h...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // SubscribeContext set context | ||||
| func SubscribeContext(ctx context.Context) SubscribeOption { | ||||
| 	return func(o *SubscribeOptions) { | ||||
|   | ||||
| @@ -17,13 +17,13 @@ func BackoffExp(_ context.Context, _ Request, attempts int) (time.Duration, erro | ||||
| } | ||||
|  | ||||
| // BackoffInterval specifies randomization interval for backoff func | ||||
| func BackoffInterval(min time.Duration, max time.Duration) BackoffFunc { | ||||
| func BackoffInterval(minTime time.Duration, maxTime time.Duration) BackoffFunc { | ||||
| 	return func(_ context.Context, _ Request, attempts int) (time.Duration, error) { | ||||
| 		td := time.Duration(math.Pow(float64(attempts), math.E)) * time.Millisecond * 100 | ||||
| 		if td < min { | ||||
| 			return min, nil | ||||
| 		} else if td > max { | ||||
| 			return max, nil | ||||
| 		if td < minTime { | ||||
| 			return minTime, nil | ||||
| 		} else if td > maxTime { | ||||
| 			return maxTime, nil | ||||
| 		} | ||||
| 		return td, nil | ||||
| 	} | ||||
|   | ||||
| @@ -34,23 +34,23 @@ func TestBackoffExp(t *testing.T) { | ||||
| } | ||||
|  | ||||
| func TestBackoffInterval(t *testing.T) { | ||||
| 	min := 100 * time.Millisecond | ||||
| 	max := 300 * time.Millisecond | ||||
| 	minTime := 100 * time.Millisecond | ||||
| 	maxTime := 300 * time.Millisecond | ||||
|  | ||||
| 	r := &testRequest{ | ||||
| 		service: "test", | ||||
| 		method:  "test", | ||||
| 	} | ||||
|  | ||||
| 	fn := BackoffInterval(min, max) | ||||
| 	fn := BackoffInterval(minTime, maxTime) | ||||
| 	for i := 0; i < 5; i++ { | ||||
| 		d, err := fn(context.TODO(), r, i) | ||||
| 		if err != nil { | ||||
| 			t.Fatal(err) | ||||
| 		} | ||||
|  | ||||
| 		if d < min || d > max { | ||||
| 			t.Fatalf("Expected %v < %v < %v", min, d, max) | ||||
| 		if d < minTime || d > maxTime { | ||||
| 			t.Fatalf("Expected %v < %v < %v", minTime, d, maxTime) | ||||
| 		} | ||||
| 	} | ||||
| } | ||||
|   | ||||
| @@ -1,5 +1,5 @@ | ||||
| // Package client is an interface for an RPC client | ||||
| package client // import "go.unistack.org/micro/v3/client" | ||||
| package client | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| @@ -44,6 +44,17 @@ type Client interface { | ||||
| 	String() string | ||||
| } | ||||
|  | ||||
| type ( | ||||
| 	FuncCall         func(ctx context.Context, req Request, rsp interface{}, opts ...CallOption) error | ||||
| 	HookCall         func(next FuncCall) FuncCall | ||||
| 	FuncStream       func(ctx context.Context, req Request, opts ...CallOption) (Stream, error) | ||||
| 	HookStream       func(next FuncStream) FuncStream | ||||
| 	FuncPublish      func(ctx context.Context, msg Message, opts ...PublishOption) error | ||||
| 	HookPublish      func(next FuncPublish) FuncPublish | ||||
| 	FuncBatchPublish func(ctx context.Context, msg []Message, opts ...PublishOption) error | ||||
| 	HookBatchPublish func(next FuncBatchPublish) FuncBatchPublish | ||||
| ) | ||||
|  | ||||
| // Message is the interface for publishing asynchronously | ||||
| type Message interface { | ||||
| 	Topic() string | ||||
|   | ||||
| @@ -1,26 +0,0 @@ | ||||
| package client | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"testing" | ||||
| 	"time" | ||||
| ) | ||||
|  | ||||
| func TestNewClientCallOptions(t *testing.T) { | ||||
| 	var flag bool | ||||
| 	w := func(fn CallFunc) CallFunc { | ||||
| 		flag = true | ||||
| 		return fn | ||||
| 	} | ||||
| 	c := NewClientCallOptions(NewClient(), | ||||
| 		WithAddress("127.0.0.1"), | ||||
| 		WithCallWrapper(w), | ||||
| 		WithRequestTimeout(1*time.Millisecond), | ||||
| 		WithRetries(0), | ||||
| 		WithBackoff(BackoffInterval(10*time.Millisecond, 100*time.Millisecond)), | ||||
| 	) | ||||
| 	_ = c.Call(context.TODO(), c.NewRequest("service", "endpoint", nil), nil) | ||||
| 	if !flag { | ||||
| 		t.Fatalf("NewClientCallOptions not works") | ||||
| 	} | ||||
| } | ||||
| @@ -15,6 +15,15 @@ func FromContext(ctx context.Context) (Client, bool) { | ||||
| 	return c, ok | ||||
| } | ||||
|  | ||||
| // MustContext get client from context | ||||
| func MustContext(ctx context.Context) Client { | ||||
| 	c, ok := FromContext(ctx) | ||||
| 	if !ok { | ||||
| 		panic("missing client") | ||||
| 	} | ||||
| 	return c | ||||
| } | ||||
|  | ||||
| // NewContext put client in context | ||||
| func NewContext(ctx context.Context, c Client) context.Context { | ||||
| 	if ctx == nil { | ||||
|   | ||||
							
								
								
									
										167
									
								
								client/noop.go
									
									
									
									
									
								
							
							
						
						
									
										167
									
								
								client/noop.go
									
									
									
									
									
								
							| @@ -4,22 +4,25 @@ import ( | ||||
| 	"context" | ||||
| 	"fmt" | ||||
| 	"os" | ||||
| 	"strconv" | ||||
| 	"time" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/broker" | ||||
| 	"go.unistack.org/micro/v3/codec" | ||||
| 	"go.unistack.org/micro/v3/errors" | ||||
| 	"go.unistack.org/micro/v3/metadata" | ||||
| 	"go.unistack.org/micro/v3/options" | ||||
| 	"go.unistack.org/micro/v3/selector" | ||||
| 	"go.unistack.org/micro/v3/semconv" | ||||
| 	"go.unistack.org/micro/v3/tracer" | ||||
| ) | ||||
|  | ||||
| // DefaultCodecs will be used to encode/decode data | ||||
| var DefaultCodecs = map[string]codec.Codec{ | ||||
| 	"application/octet-stream": codec.NewCodec(), | ||||
| } | ||||
|  | ||||
| type noopClient struct { | ||||
| 	opts Options | ||||
| 	funcPublish      FuncPublish | ||||
| 	funcBatchPublish FuncBatchPublish | ||||
| 	funcCall         FuncCall | ||||
| 	funcStream       FuncStream | ||||
| 	opts             Options | ||||
| } | ||||
|  | ||||
| type noopMessage struct { | ||||
| @@ -40,16 +43,14 @@ type noopRequest struct { | ||||
|  | ||||
| // NewClient returns new noop client | ||||
| func NewClient(opts ...Option) Client { | ||||
| 	nc := &noopClient{opts: NewOptions(opts...)} | ||||
| 	// wrap in reverse | ||||
| 	n := &noopClient{opts: NewOptions(opts...)} | ||||
|  | ||||
| 	c := Client(nc) | ||||
| 	n.funcCall = n.fnCall | ||||
| 	n.funcStream = n.fnStream | ||||
| 	n.funcPublish = n.fnPublish | ||||
| 	n.funcBatchPublish = n.fnBatchPublish | ||||
|  | ||||
| 	for i := len(nc.opts.Wrappers); i > 0; i-- { | ||||
| 		c = nc.opts.Wrappers[i-1](c) | ||||
| 	} | ||||
|  | ||||
| 	return c | ||||
| 	return n | ||||
| } | ||||
|  | ||||
| func (n *noopClient) Name() string { | ||||
| @@ -101,10 +102,13 @@ func (n *noopResponse) Read() ([]byte, error) { | ||||
| 	return nil, nil | ||||
| } | ||||
|  | ||||
| type noopStream struct{} | ||||
| type noopStream struct { | ||||
| 	err error | ||||
| 	ctx context.Context | ||||
| } | ||||
|  | ||||
| func (n *noopStream) Context() context.Context { | ||||
| 	return context.Background() | ||||
| 	return n.ctx | ||||
| } | ||||
|  | ||||
| func (n *noopStream) Request() Request { | ||||
| @@ -132,15 +136,21 @@ func (n *noopStream) RecvMsg(interface{}) error { | ||||
| } | ||||
|  | ||||
| func (n *noopStream) Error() error { | ||||
| 	return nil | ||||
| 	return n.err | ||||
| } | ||||
|  | ||||
| func (n *noopStream) Close() error { | ||||
| 	return nil | ||||
| 	if sp, ok := tracer.SpanFromContext(n.ctx); ok && sp != nil { | ||||
| 		if n.err != nil { | ||||
| 			sp.SetStatus(tracer.SpanStatusError, n.err.Error()) | ||||
| 		} | ||||
| 		sp.Finish() | ||||
| 	} | ||||
| 	return n.err | ||||
| } | ||||
|  | ||||
| func (n *noopStream) CloseSend() error { | ||||
| 	return nil | ||||
| 	return n.err | ||||
| } | ||||
|  | ||||
| func (n *noopMessage) Topic() string { | ||||
| @@ -163,9 +173,6 @@ func (n *noopClient) newCodec(contentType string) (codec.Codec, error) { | ||||
| 	if cf, ok := n.opts.Codecs[contentType]; ok { | ||||
| 		return cf, nil | ||||
| 	} | ||||
| 	if cf, ok := DefaultCodecs[contentType]; ok { | ||||
| 		return cf, nil | ||||
| 	} | ||||
| 	return nil, codec.ErrUnknownContentType | ||||
| } | ||||
|  | ||||
| @@ -173,6 +180,25 @@ func (n *noopClient) Init(opts ...Option) error { | ||||
| 	for _, o := range opts { | ||||
| 		o(&n.opts) | ||||
| 	} | ||||
|  | ||||
| 	n.funcCall = n.fnCall | ||||
| 	n.funcStream = n.fnStream | ||||
| 	n.funcPublish = n.fnPublish | ||||
| 	n.funcBatchPublish = n.fnBatchPublish | ||||
|  | ||||
| 	n.opts.Hooks.EachPrev(func(hook options.Hook) { | ||||
| 		switch h := hook.(type) { | ||||
| 		case HookCall: | ||||
| 			n.funcCall = h(n.funcCall) | ||||
| 		case HookStream: | ||||
| 			n.funcStream = h(n.funcStream) | ||||
| 		case HookPublish: | ||||
| 			n.funcPublish = h(n.funcPublish) | ||||
| 		case HookBatchPublish: | ||||
| 			n.funcBatchPublish = h(n.funcBatchPublish) | ||||
| 		} | ||||
| 	}) | ||||
|  | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| @@ -185,6 +211,31 @@ func (n *noopClient) String() string { | ||||
| } | ||||
|  | ||||
| func (n *noopClient) Call(ctx context.Context, req Request, rsp interface{}, opts ...CallOption) error { | ||||
| 	ts := time.Now() | ||||
| 	n.opts.Meter.Counter(semconv.ClientRequestInflight, "endpoint", req.Endpoint()).Inc() | ||||
| 	var sp tracer.Span | ||||
| 	ctx, sp = n.opts.Tracer.Start(ctx, "rpc-client", | ||||
| 		tracer.WithSpanKind(tracer.SpanKindClient), | ||||
| 		tracer.WithSpanLabels("endpoint", req.Endpoint()), | ||||
| 	) | ||||
| 	err := n.funcCall(ctx, req, rsp, opts...) | ||||
| 	n.opts.Meter.Counter(semconv.ClientRequestInflight, "endpoint", req.Endpoint()).Dec() | ||||
| 	te := time.Since(ts) | ||||
| 	n.opts.Meter.Summary(semconv.ClientRequestLatencyMicroseconds, "endpoint", req.Endpoint()).Update(te.Seconds()) | ||||
| 	n.opts.Meter.Histogram(semconv.ClientRequestDurationSeconds, "endpoint", req.Endpoint()).Update(te.Seconds()) | ||||
|  | ||||
| 	if me := errors.FromError(err); me == nil { | ||||
| 		sp.Finish() | ||||
| 		n.opts.Meter.Counter(semconv.ClientRequestTotal, "endpoint", req.Endpoint(), "status", "success", "code", strconv.Itoa(int(200))).Inc() | ||||
| 	} else { | ||||
| 		sp.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 		n.opts.Meter.Counter(semconv.ClientRequestTotal, "endpoint", req.Endpoint(), "status", "failure", "code", strconv.Itoa(int(me.Code))).Inc() | ||||
| 	} | ||||
|  | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| func (n *noopClient) fnCall(ctx context.Context, req Request, rsp interface{}, opts ...CallOption) error { | ||||
| 	// make a copy of call opts | ||||
| 	callOpts := n.opts.CallOptions | ||||
| 	for _, opt := range opts { | ||||
| @@ -213,11 +264,8 @@ func (n *noopClient) Call(ctx context.Context, req Request, rsp interface{}, opt | ||||
| 	} | ||||
|  | ||||
| 	// make copy of call method | ||||
| 	hcall := n.call | ||||
|  | ||||
| 	// wrap the call in reverse | ||||
| 	for i := len(callOpts.CallWrappers); i > 0; i-- { | ||||
| 		hcall = callOpts.CallWrappers[i-1](hcall) | ||||
| 	hcall := func(ctx context.Context, addr string, req Request, rsp interface{}, opts CallOptions) error { | ||||
| 		return nil | ||||
| 	} | ||||
|  | ||||
| 	// use the router passed as a call option, or fallback to the rpc clients router | ||||
| @@ -242,7 +290,7 @@ func (n *noopClient) Call(ctx context.Context, req Request, rsp interface{}, opt | ||||
| 		// call backoff first. Someone may want an initial start delay | ||||
| 		t, err := callOpts.Backoff(ctx, req, i) | ||||
| 		if err != nil { | ||||
| 			return errors.InternalServerError("go.micro.client", err.Error()) | ||||
| 			return errors.InternalServerError("go.micro.client", "%s", err) | ||||
| 		} | ||||
|  | ||||
| 		// only sleep if greater than 0 | ||||
| @@ -256,7 +304,7 @@ func (n *noopClient) Call(ctx context.Context, req Request, rsp interface{}, opt | ||||
| 			// TODO apply any filtering here | ||||
| 			routes, err = n.opts.Lookup(ctx, req, callOpts) | ||||
| 			if err != nil { | ||||
| 				return errors.InternalServerError("go.micro.client", err.Error()) | ||||
| 				return errors.InternalServerError("go.micro.client", "%s", err) | ||||
| 			} | ||||
|  | ||||
| 			// balance the list of nodes | ||||
| @@ -316,11 +364,7 @@ func (n *noopClient) Call(ctx context.Context, req Request, rsp interface{}, opt | ||||
| 	return gerr | ||||
| } | ||||
|  | ||||
| func (n *noopClient) call(ctx context.Context, addr string, req Request, rsp interface{}, opts CallOptions) error { | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| func (n *noopClient) NewRequest(service, endpoint string, req interface{}, opts ...RequestOption) Request { | ||||
| func (n *noopClient) NewRequest(service, endpoint string, _ interface{}, _ ...RequestOption) Request { | ||||
| 	return &noopRequest{service: service, endpoint: endpoint} | ||||
| } | ||||
|  | ||||
| @@ -330,6 +374,31 @@ func (n *noopClient) NewMessage(topic string, msg interface{}, opts ...MessageOp | ||||
| } | ||||
|  | ||||
| func (n *noopClient) Stream(ctx context.Context, req Request, opts ...CallOption) (Stream, error) { | ||||
| 	ts := time.Now() | ||||
| 	n.opts.Meter.Counter(semconv.ClientRequestInflight, "endpoint", req.Endpoint()).Inc() | ||||
| 	var sp tracer.Span | ||||
| 	ctx, sp = n.opts.Tracer.Start(ctx, "rpc-client", | ||||
| 		tracer.WithSpanKind(tracer.SpanKindClient), | ||||
| 		tracer.WithSpanLabels("endpoint", req.Endpoint()), | ||||
| 	) | ||||
| 	stream, err := n.funcStream(ctx, req, opts...) | ||||
| 	n.opts.Meter.Counter(semconv.ClientRequestInflight, "endpoint", req.Endpoint()).Dec() | ||||
| 	te := time.Since(ts) | ||||
| 	n.opts.Meter.Summary(semconv.ClientRequestLatencyMicroseconds, "endpoint", req.Endpoint()).Update(te.Seconds()) | ||||
| 	n.opts.Meter.Histogram(semconv.ClientRequestDurationSeconds, "endpoint", req.Endpoint()).Update(te.Seconds()) | ||||
|  | ||||
| 	if me := errors.FromError(err); me == nil { | ||||
| 		sp.Finish() | ||||
| 		n.opts.Meter.Counter(semconv.ClientRequestTotal, "endpoint", req.Endpoint(), "status", "success", "code", strconv.Itoa(int(200))).Inc() | ||||
| 	} else { | ||||
| 		sp.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 		n.opts.Meter.Counter(semconv.ClientRequestTotal, "endpoint", req.Endpoint(), "status", "failure", "code", strconv.Itoa(int(me.Code))).Inc() | ||||
| 	} | ||||
|  | ||||
| 	return stream, err | ||||
| } | ||||
|  | ||||
| func (n *noopClient) fnStream(ctx context.Context, req Request, opts ...CallOption) (Stream, error) { | ||||
| 	var err error | ||||
|  | ||||
| 	// make a copy of call opts | ||||
| @@ -389,7 +458,7 @@ func (n *noopClient) Stream(ctx context.Context, req Request, opts ...CallOption | ||||
| 		// call backoff first. Someone may want an initial start delay | ||||
| 		t, cerr := callOpts.Backoff(ctx, req, i) | ||||
| 		if cerr != nil { | ||||
| 			return nil, errors.InternalServerError("go.micro.client", cerr.Error()) | ||||
| 			return nil, errors.InternalServerError("go.micro.client", "%s", cerr) | ||||
| 		} | ||||
|  | ||||
| 		// only sleep if greater than 0 | ||||
| @@ -403,7 +472,7 @@ func (n *noopClient) Stream(ctx context.Context, req Request, opts ...CallOption | ||||
| 			// TODO apply any filtering here | ||||
| 			routes, err = n.opts.Lookup(ctx, req, callOpts) | ||||
| 			if err != nil { | ||||
| 				return nil, errors.InternalServerError("go.micro.client", err.Error()) | ||||
| 				return nil, errors.InternalServerError("go.micro.client", "%s", err) | ||||
| 			} | ||||
|  | ||||
| 			// balance the list of nodes | ||||
| @@ -469,15 +538,23 @@ func (n *noopClient) Stream(ctx context.Context, req Request, opts ...CallOption | ||||
| 	return nil, grr | ||||
| } | ||||
|  | ||||
| func (n *noopClient) stream(ctx context.Context, addr string, req Request, opts CallOptions) (Stream, error) { | ||||
| 	return &noopStream{}, nil | ||||
| func (n *noopClient) stream(ctx context.Context, _ string, _ Request, _ CallOptions) (Stream, error) { | ||||
| 	return &noopStream{ctx: ctx}, nil | ||||
| } | ||||
|  | ||||
| func (n *noopClient) BatchPublish(ctx context.Context, ps []Message, opts ...PublishOption) error { | ||||
| 	return n.funcBatchPublish(ctx, ps, opts...) | ||||
| } | ||||
|  | ||||
| func (n *noopClient) fnBatchPublish(ctx context.Context, ps []Message, opts ...PublishOption) error { | ||||
| 	return n.publish(ctx, ps, opts...) | ||||
| } | ||||
|  | ||||
| func (n *noopClient) Publish(ctx context.Context, p Message, opts ...PublishOption) error { | ||||
| 	return n.funcPublish(ctx, p, opts...) | ||||
| } | ||||
|  | ||||
| func (n *noopClient) fnPublish(ctx context.Context, p Message, opts ...PublishOption) error { | ||||
| 	return n.publish(ctx, []Message{p}, opts...) | ||||
| } | ||||
|  | ||||
| @@ -503,7 +580,6 @@ func (n *noopClient) publish(ctx context.Context, ps []Message, opts ...PublishO | ||||
|  | ||||
| 	for _, p := range ps { | ||||
| 		md := metadata.Copy(omd) | ||||
| 		md[metadata.HeaderContentType] = p.ContentType() | ||||
| 		topic := p.Topic() | ||||
| 		if len(exchange) > 0 { | ||||
| 			topic = exchange | ||||
| @@ -515,6 +591,8 @@ func (n *noopClient) publish(ctx context.Context, ps []Message, opts ...PublishO | ||||
| 			md.Set(k, v) | ||||
| 		} | ||||
|  | ||||
| 		md[metadata.HeaderContentType] = p.ContentType() | ||||
|  | ||||
| 		var body []byte | ||||
|  | ||||
| 		// passed in raw data | ||||
| @@ -524,13 +602,13 @@ func (n *noopClient) publish(ctx context.Context, ps []Message, opts ...PublishO | ||||
| 			// use codec for payload | ||||
| 			cf, err := n.newCodec(p.ContentType()) | ||||
| 			if err != nil { | ||||
| 				return errors.InternalServerError("go.micro.client", err.Error()) | ||||
| 				return errors.InternalServerError("go.micro.client", "%s", err) | ||||
| 			} | ||||
|  | ||||
| 			// set the body | ||||
| 			b, err := cf.Marshal(p.Payload()) | ||||
| 			if err != nil { | ||||
| 				return errors.InternalServerError("go.micro.client", err.Error()) | ||||
| 				return errors.InternalServerError("go.micro.client", "%s", err) | ||||
| 			} | ||||
| 			body = b | ||||
| 		} | ||||
| @@ -538,6 +616,13 @@ func (n *noopClient) publish(ctx context.Context, ps []Message, opts ...PublishO | ||||
| 		msgs = append(msgs, &broker.Message{Header: md, Body: body}) | ||||
| 	} | ||||
|  | ||||
| 	if len(msgs) == 1 { | ||||
| 		return n.opts.Broker.Publish(ctx, msgs[0].Header[metadata.HeaderTopic], msgs[0], | ||||
| 			broker.PublishContext(options.Context), | ||||
| 			broker.PublishBodyOnly(options.BodyOnly), | ||||
| 		) | ||||
| 	} | ||||
|  | ||||
| 	return n.opts.Broker.BatchPublish(ctx, msgs, | ||||
| 		broker.PublishContext(options.Context), | ||||
| 		broker.PublishBodyOnly(options.BodyOnly), | ||||
|   | ||||
							
								
								
									
										37
									
								
								client/noop_test.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										37
									
								
								client/noop_test.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,37 @@ | ||||
| package client | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"testing" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/codec" | ||||
| ) | ||||
|  | ||||
| type testHook struct { | ||||
| 	f bool | ||||
| } | ||||
|  | ||||
| func (t *testHook) Publish(fn FuncPublish) FuncPublish { | ||||
| 	return func(ctx context.Context, msg Message, opts ...PublishOption) error { | ||||
| 		t.f = true | ||||
| 		return fn(ctx, msg, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestNoopHook(t *testing.T) { | ||||
| 	h := &testHook{} | ||||
|  | ||||
| 	c := NewClient(Codec("application/octet-stream", codec.NewCodec()), Hooks(HookPublish(h.Publish))) | ||||
|  | ||||
| 	if err := c.Init(); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	if err := c.Publish(context.TODO(), c.NewMessage("", nil, MessageContentType("application/octet-stream"))); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	if !h.f { | ||||
| 		t.Fatal("hook not works") | ||||
| 	} | ||||
| } | ||||
| @@ -11,7 +11,7 @@ import ( | ||||
| 	"go.unistack.org/micro/v3/logger" | ||||
| 	"go.unistack.org/micro/v3/metadata" | ||||
| 	"go.unistack.org/micro/v3/meter" | ||||
| 	"go.unistack.org/micro/v3/network/transport" | ||||
| 	"go.unistack.org/micro/v3/options" | ||||
| 	"go.unistack.org/micro/v3/register" | ||||
| 	"go.unistack.org/micro/v3/router" | ||||
| 	"go.unistack.org/micro/v3/selector" | ||||
| @@ -21,8 +21,16 @@ import ( | ||||
|  | ||||
| // Options holds client options | ||||
| type Options struct { | ||||
| 	// Transport used for transfer messages | ||||
| 	Transport transport.Transport | ||||
| 	// Codecs map | ||||
| 	Codecs map[string]codec.Codec | ||||
|  | ||||
| 	// Proxy is used for proxy requests | ||||
| 	Proxy string | ||||
| 	// ContentType is used to select codec | ||||
| 	ContentType string | ||||
| 	// Name is the client name | ||||
| 	Name string | ||||
|  | ||||
| 	// Selector used to select needed address | ||||
| 	Selector selector.Selector | ||||
| 	// Logger used to log messages | ||||
| @@ -37,28 +45,28 @@ type Options struct { | ||||
| 	Context context.Context | ||||
| 	// Router used to get route | ||||
| 	Router router.Router | ||||
|  | ||||
| 	// TLSConfig specifies tls.Config for secure connection | ||||
| 	TLSConfig *tls.Config | ||||
| 	// Codecs map | ||||
| 	Codecs map[string]codec.Codec | ||||
|  | ||||
| 	// Lookup func used to get destination addr | ||||
| 	Lookup LookupFunc | ||||
| 	// Proxy is used for proxy requests | ||||
| 	Proxy string | ||||
| 	// ContentType is used to select codec | ||||
| 	ContentType string | ||||
| 	// Name is the client name | ||||
| 	Name string | ||||
| 	// ContextDialer used to connect | ||||
| 	ContextDialer func(context.Context, string) (net.Conn, error) | ||||
|  | ||||
| 	// Wrappers contains wrappers | ||||
| 	Wrappers []Wrapper | ||||
| 	// Hooks can be run before broker Publish/BatchPublish and | ||||
| 	// Subscribe/BatchSubscribe methods | ||||
| 	Hooks options.Hooks | ||||
|  | ||||
| 	// CallOptions contains default CallOptions | ||||
| 	CallOptions CallOptions | ||||
|  | ||||
| 	// PoolSize connection pool size | ||||
| 	PoolSize int | ||||
| 	// PoolTTL connection pool ttl | ||||
| 	PoolTTL time.Duration | ||||
| 	// ContextDialer used to connect | ||||
| 	ContextDialer func(context.Context, string) (net.Conn, error) | ||||
| } | ||||
|  | ||||
| // NewCallOptions creates new call options struct | ||||
| @@ -72,6 +80,16 @@ func NewCallOptions(opts ...CallOption) CallOptions { | ||||
|  | ||||
| // CallOptions holds client call options | ||||
| type CallOptions struct { | ||||
| 	// RequestMetadata holds additional metadata for call | ||||
| 	RequestMetadata metadata.Metadata | ||||
|  | ||||
| 	// Network name | ||||
| 	Network string | ||||
| 	// Content-Type | ||||
| 	ContentType string | ||||
| 	// AuthToken string | ||||
| 	AuthToken string | ||||
|  | ||||
| 	// Selector selects addr | ||||
| 	Selector selector.Selector | ||||
| 	// Context used for deadline | ||||
| @@ -79,35 +97,30 @@ type CallOptions struct { | ||||
| 	// Router used for route | ||||
| 	Router router.Router | ||||
| 	// Retry func used for retries | ||||
|  | ||||
| 	// ResponseMetadata holds additional metadata from call | ||||
| 	ResponseMetadata *metadata.Metadata | ||||
|  | ||||
| 	Retry RetryFunc | ||||
| 	// Backoff func used for backoff when retry | ||||
| 	Backoff BackoffFunc | ||||
| 	// Network name | ||||
| 	Network string | ||||
| 	// Content-Type | ||||
| 	ContentType string | ||||
| 	// AuthToken string | ||||
| 	AuthToken string | ||||
| 	// ContextDialer used to connect | ||||
| 	ContextDialer func(context.Context, string) (net.Conn, error) | ||||
|  | ||||
| 	// Address specifies static addr list | ||||
| 	Address []string | ||||
| 	// SelectOptions selector options | ||||
| 	SelectOptions []selector.SelectOption | ||||
| 	// CallWrappers call wrappers | ||||
| 	CallWrappers []CallWrapper | ||||
|  | ||||
| 	// StreamTimeout stream timeout | ||||
| 	StreamTimeout time.Duration | ||||
| 	// RequestTimeout request timeout | ||||
| 	RequestTimeout time.Duration | ||||
| 	// RequestMetadata holds additional metadata for call | ||||
| 	RequestMetadata metadata.Metadata | ||||
| 	// ResponseMetadata holds additional metadata from call | ||||
| 	ResponseMetadata *metadata.Metadata | ||||
|  | ||||
| 	// DialTimeout dial timeout | ||||
| 	DialTimeout time.Duration | ||||
| 	// Retries specifies retries num | ||||
| 	Retries int | ||||
| 	// ContextDialer used to connect | ||||
| 	ContextDialer func(context.Context, string) (net.Conn, error) | ||||
| } | ||||
|  | ||||
| // ContextDialer pass ContextDialer to client | ||||
| @@ -192,18 +205,16 @@ func NewOptions(opts ...Option) Options { | ||||
| 			Retry:          DefaultRetry, | ||||
| 			Retries:        DefaultRetries, | ||||
| 			RequestTimeout: DefaultRequestTimeout, | ||||
| 			DialTimeout:    transport.DefaultDialTimeout, | ||||
| 		}, | ||||
| 		Lookup:    LookupRoute, | ||||
| 		PoolSize:  DefaultPoolSize, | ||||
| 		PoolTTL:   DefaultPoolTTL, | ||||
| 		Selector:  random.NewSelector(), | ||||
| 		Logger:    logger.DefaultLogger, | ||||
| 		Broker:    broker.DefaultBroker, | ||||
| 		Meter:     meter.DefaultMeter, | ||||
| 		Tracer:    tracer.DefaultTracer, | ||||
| 		Router:    router.DefaultRouter, | ||||
| 		Transport: transport.DefaultTransport, | ||||
| 		Lookup:   LookupRoute, | ||||
| 		PoolSize: DefaultPoolSize, | ||||
| 		PoolTTL:  DefaultPoolTTL, | ||||
| 		Selector: random.NewSelector(), | ||||
| 		Logger:   logger.DefaultLogger, | ||||
| 		Broker:   broker.DefaultBroker, | ||||
| 		Meter:    meter.DefaultMeter, | ||||
| 		Tracer:   tracer.DefaultTracer, | ||||
| 		Router:   router.DefaultRouter, | ||||
| 	} | ||||
|  | ||||
| 	for _, o := range opts { | ||||
| @@ -276,13 +287,6 @@ func PoolTTL(d time.Duration) Option { | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Transport to use for communication e.g http, rabbitmq, etc | ||||
| func Transport(t transport.Transport) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Transport = t | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Register sets the routers register | ||||
| func Register(r register.Register) Option { | ||||
| 	return func(o *Options) { | ||||
| @@ -306,20 +310,6 @@ func Selector(s selector.Selector) Option { | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Wrap adds a wrapper to the list of options passed into the client | ||||
| func Wrap(w Wrapper) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Wrappers = append(o.Wrappers, w) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WrapCall adds a wrapper to the list of CallFunc wrappers | ||||
| func WrapCall(cw ...CallWrapper) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.CallOptions.CallWrappers = append(o.CallOptions.CallWrappers, cw...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Backoff is used to set the backoff function used when retrying Calls | ||||
| func Backoff(fn BackoffFunc) Option { | ||||
| 	return func(o *Options) { | ||||
| @@ -346,14 +336,6 @@ func TLSConfig(t *tls.Config) Option { | ||||
| 	return func(o *Options) { | ||||
| 		// set the internal tls | ||||
| 		o.TLSConfig = t | ||||
|  | ||||
| 		// set the default transport if one is not | ||||
| 		// already set. Required for Init call below. | ||||
|  | ||||
| 		// set the transport tls | ||||
| 		_ = o.Transport.Init( | ||||
| 			transport.TLSConfig(t), | ||||
| 		) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| @@ -450,13 +432,6 @@ func WithAddress(a ...string) CallOption { | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithCallWrapper is a CallOption which adds to the existing CallFunc wrappers | ||||
| func WithCallWrapper(cw ...CallWrapper) CallOption { | ||||
| 	return func(o *CallOptions) { | ||||
| 		o.CallWrappers = append(o.CallWrappers, cw...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithBackoff is a CallOption which overrides that which | ||||
| // set in Options.CallOptions | ||||
| func WithBackoff(fn BackoffFunc) CallOption { | ||||
| @@ -526,13 +501,6 @@ func WithAuthToken(t string) CallOption { | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithNetwork is a CallOption which sets the network attribute | ||||
| func WithNetwork(n string) CallOption { | ||||
| 	return func(o *CallOptions) { | ||||
| 		o.Network = n | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithRouter sets the router to use for this call | ||||
| func WithRouter(r router.Router) CallOption { | ||||
| 	return func(o *CallOptions) { | ||||
| @@ -591,3 +559,10 @@ func RequestContentType(ct string) RequestOption { | ||||
| 		o.ContentType = ct | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Hooks sets hook runs before action | ||||
| func Hooks(h ...options.Hook) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Hooks = append(o.Hooks, h...) | ||||
| 	} | ||||
| } | ||||
|   | ||||
| @@ -38,4 +38,10 @@ type Cluster interface { | ||||
| 	Broadcast(ctx context.Context, msg Message, filter ...string) error | ||||
| 	// Unicast send message to single member in cluster | ||||
| 	Unicast(ctx context.Context, node Node, msg Message) error | ||||
| 	// Live returns cluster liveness | ||||
| 	Live() bool | ||||
| 	// Ready returns cluster readiness | ||||
| 	Ready() bool | ||||
| 	// Health returns cluster health | ||||
| 	Health() bool | ||||
| } | ||||
|   | ||||
							
								
								
									
										235
									
								
								cluster/hasql/cluster.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										235
									
								
								cluster/hasql/cluster.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,235 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"database/sql" | ||||
| 	"reflect" | ||||
| 	"unsafe" | ||||
|  | ||||
| 	"golang.yandex/hasql/v2" | ||||
| ) | ||||
|  | ||||
| func newSQLRowError() *sql.Row { | ||||
| 	row := &sql.Row{} | ||||
| 	t := reflect.TypeOf(row).Elem() | ||||
| 	field, _ := t.FieldByName("err") | ||||
| 	rowPtr := unsafe.Pointer(row) | ||||
| 	errFieldPtr := unsafe.Pointer(uintptr(rowPtr) + field.Offset) | ||||
| 	errPtr := (*error)(errFieldPtr) | ||||
| 	*errPtr = ErrorNoAliveNodes | ||||
| 	return row | ||||
| } | ||||
|  | ||||
| type ClusterQuerier interface { | ||||
| 	Querier | ||||
| 	WaitForNodes(ctx context.Context, criterion ...hasql.NodeStateCriterion) error | ||||
| } | ||||
|  | ||||
| type Cluster struct { | ||||
| 	hasql   *hasql.Cluster[Querier] | ||||
| 	options ClusterOptions | ||||
| } | ||||
|  | ||||
| // NewCluster returns [Querier] that provides cluster of nodes | ||||
| func NewCluster[T Querier](opts ...ClusterOption) (ClusterQuerier, error) { | ||||
| 	options := ClusterOptions{Context: context.Background()} | ||||
| 	for _, opt := range opts { | ||||
| 		opt(&options) | ||||
| 	} | ||||
| 	if options.NodeChecker == nil { | ||||
| 		return nil, ErrClusterChecker | ||||
| 	} | ||||
| 	if options.NodeDiscoverer == nil { | ||||
| 		return nil, ErrClusterDiscoverer | ||||
| 	} | ||||
| 	if options.NodePicker == nil { | ||||
| 		return nil, ErrClusterPicker | ||||
| 	} | ||||
|  | ||||
| 	if options.Retries < 1 { | ||||
| 		options.Retries = 1 | ||||
| 	} | ||||
|  | ||||
| 	if options.NodeStateCriterion == 0 { | ||||
| 		options.NodeStateCriterion = hasql.Primary | ||||
| 	} | ||||
|  | ||||
| 	options.Options = append(options.Options, hasql.WithNodePicker(options.NodePicker)) | ||||
| 	if p, ok := options.NodePicker.(*CustomPicker[Querier]); ok { | ||||
| 		p.opts.Priority = options.NodePriority | ||||
| 	} | ||||
|  | ||||
| 	c, err := hasql.NewCluster( | ||||
| 		options.NodeDiscoverer, | ||||
| 		options.NodeChecker, | ||||
| 		options.Options..., | ||||
| 	) | ||||
| 	if err != nil { | ||||
| 		return nil, err | ||||
| 	} | ||||
|  | ||||
| 	return &Cluster{hasql: c, options: options}, nil | ||||
| } | ||||
|  | ||||
| func (c *Cluster) BeginTx(ctx context.Context, opts *sql.TxOptions) (*sql.Tx, error) { | ||||
| 	var tx *sql.Tx | ||||
| 	var err error | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			if tx, err = n.DB().BeginTx(ctx, opts); err != nil && retries >= c.options.Retries { | ||||
| 				return true | ||||
| 			} | ||||
| 		} | ||||
| 		return false | ||||
| 	}) | ||||
|  | ||||
| 	if tx == nil && err == nil { | ||||
| 		err = ErrorNoAliveNodes | ||||
| 	} | ||||
|  | ||||
| 	return tx, err | ||||
| } | ||||
|  | ||||
| func (c *Cluster) Close() error { | ||||
| 	return c.hasql.Close() | ||||
| } | ||||
|  | ||||
| func (c *Cluster) Conn(ctx context.Context) (*sql.Conn, error) { | ||||
| 	var conn *sql.Conn | ||||
| 	var err error | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			if conn, err = n.DB().Conn(ctx); err != nil && retries >= c.options.Retries { | ||||
| 				return true | ||||
| 			} | ||||
| 		} | ||||
| 		return false | ||||
| 	}) | ||||
|  | ||||
| 	if conn == nil && err == nil { | ||||
| 		err = ErrorNoAliveNodes | ||||
| 	} | ||||
|  | ||||
| 	return conn, err | ||||
| } | ||||
|  | ||||
| func (c *Cluster) ExecContext(ctx context.Context, query string, args ...interface{}) (sql.Result, error) { | ||||
| 	var res sql.Result | ||||
| 	var err error | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			if res, err = n.DB().ExecContext(ctx, query, args...); err != nil && retries >= c.options.Retries { | ||||
| 				return true | ||||
| 			} | ||||
| 		} | ||||
| 		return false | ||||
| 	}) | ||||
|  | ||||
| 	if res == nil && err == nil { | ||||
| 		err = ErrorNoAliveNodes | ||||
| 	} | ||||
|  | ||||
| 	return res, err | ||||
| } | ||||
|  | ||||
| func (c *Cluster) PrepareContext(ctx context.Context, query string) (*sql.Stmt, error) { | ||||
| 	var res *sql.Stmt | ||||
| 	var err error | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			if res, err = n.DB().PrepareContext(ctx, query); err != nil && retries >= c.options.Retries { | ||||
| 				return true | ||||
| 			} | ||||
| 		} | ||||
| 		return false | ||||
| 	}) | ||||
|  | ||||
| 	if res == nil && err == nil { | ||||
| 		err = ErrorNoAliveNodes | ||||
| 	} | ||||
|  | ||||
| 	return res, err | ||||
| } | ||||
|  | ||||
| func (c *Cluster) QueryContext(ctx context.Context, query string, args ...interface{}) (*sql.Rows, error) { | ||||
| 	var res *sql.Rows | ||||
| 	var err error | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			if res, err = n.DB().QueryContext(ctx, query); err != nil && err != sql.ErrNoRows && retries >= c.options.Retries { | ||||
| 				return true | ||||
| 			} | ||||
| 		} | ||||
| 		return false | ||||
| 	}) | ||||
|  | ||||
| 	if res == nil && err == nil { | ||||
| 		err = ErrorNoAliveNodes | ||||
| 	} | ||||
|  | ||||
| 	return res, err | ||||
| } | ||||
|  | ||||
| func (c *Cluster) QueryRowContext(ctx context.Context, query string, args ...interface{}) *sql.Row { | ||||
| 	var res *sql.Row | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			res = n.DB().QueryRowContext(ctx, query, args...) | ||||
| 			if res.Err() == nil { | ||||
| 				return false | ||||
| 			} else if res.Err() != nil && retries >= c.options.Retries { | ||||
| 				return false | ||||
| 			} | ||||
| 		} | ||||
| 		return true | ||||
| 	}) | ||||
|  | ||||
| 	if res == nil { | ||||
| 		res = newSQLRowError() | ||||
| 	} | ||||
|  | ||||
| 	return res | ||||
| } | ||||
|  | ||||
| func (c *Cluster) PingContext(ctx context.Context) error { | ||||
| 	var err error | ||||
| 	var ok bool | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		ok = true | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			if err = n.DB().PingContext(ctx); err != nil && retries >= c.options.Retries { | ||||
| 				return true | ||||
| 			} | ||||
| 		} | ||||
| 		return false | ||||
| 	}) | ||||
|  | ||||
| 	if !ok { | ||||
| 		err = ErrorNoAliveNodes | ||||
| 	} | ||||
|  | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| func (c *Cluster) WaitForNodes(ctx context.Context, criterions ...hasql.NodeStateCriterion) error { | ||||
| 	for _, criterion := range criterions { | ||||
| 		if _, err := c.hasql.WaitForNode(ctx, criterion); err != nil { | ||||
| 			return err | ||||
| 		} | ||||
| 	} | ||||
| 	return nil | ||||
| } | ||||
							
								
								
									
										171
									
								
								cluster/hasql/cluster_test.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										171
									
								
								cluster/hasql/cluster_test.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,171 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"fmt" | ||||
| 	"testing" | ||||
| 	"time" | ||||
|  | ||||
| 	"github.com/DATA-DOG/go-sqlmock" | ||||
| 	"golang.yandex/hasql/v2" | ||||
| ) | ||||
|  | ||||
| func TestNewCluster(t *testing.T) { | ||||
| 	dbMaster, dbMasterMock, err := sqlmock.New(sqlmock.MonitorPingsOption(true)) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer dbMaster.Close() | ||||
| 	dbMasterMock.MatchExpectationsInOrder(false) | ||||
|  | ||||
| 	dbMasterMock.ExpectQuery(`.*pg_is_in_recovery.*`).WillReturnRows( | ||||
| 		sqlmock.NewRowsWithColumnDefinition( | ||||
| 			sqlmock.NewColumn("role").OfType("int8", 0), | ||||
| 			sqlmock.NewColumn("replication_lag").OfType("int8", 0)). | ||||
| 			AddRow(1, 0)). | ||||
| 		RowsWillBeClosed(). | ||||
| 		WithoutArgs() | ||||
|  | ||||
| 	dbMasterMock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("master-dc1")) | ||||
|  | ||||
| 	dbDRMaster, dbDRMasterMock, err := sqlmock.New(sqlmock.MonitorPingsOption(true)) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer dbDRMaster.Close() | ||||
| 	dbDRMasterMock.MatchExpectationsInOrder(false) | ||||
|  | ||||
| 	dbDRMasterMock.ExpectQuery(`.*pg_is_in_recovery.*`).WillReturnRows( | ||||
| 		sqlmock.NewRowsWithColumnDefinition( | ||||
| 			sqlmock.NewColumn("role").OfType("int8", 0), | ||||
| 			sqlmock.NewColumn("replication_lag").OfType("int8", 0)). | ||||
| 			AddRow(2, 40)). | ||||
| 		RowsWillBeClosed(). | ||||
| 		WithoutArgs() | ||||
|  | ||||
| 	dbDRMasterMock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("drmaster1-dc2")) | ||||
|  | ||||
| 	dbDRMasterMock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("drmaster")) | ||||
|  | ||||
| 	dbSlaveDC1, dbSlaveDC1Mock, err := sqlmock.New(sqlmock.MonitorPingsOption(true)) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer dbSlaveDC1.Close() | ||||
| 	dbSlaveDC1Mock.MatchExpectationsInOrder(false) | ||||
|  | ||||
| 	dbSlaveDC1Mock.ExpectQuery(`.*pg_is_in_recovery.*`).WillReturnRows( | ||||
| 		sqlmock.NewRowsWithColumnDefinition( | ||||
| 			sqlmock.NewColumn("role").OfType("int8", 0), | ||||
| 			sqlmock.NewColumn("replication_lag").OfType("int8", 0)). | ||||
| 			AddRow(2, 50)). | ||||
| 		RowsWillBeClosed(). | ||||
| 		WithoutArgs() | ||||
|  | ||||
| 	dbSlaveDC1Mock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("slave-dc1")) | ||||
|  | ||||
| 	dbSlaveDC2, dbSlaveDC2Mock, err := sqlmock.New(sqlmock.MonitorPingsOption(true)) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer dbSlaveDC2.Close() | ||||
| 	dbSlaveDC1Mock.MatchExpectationsInOrder(false) | ||||
|  | ||||
| 	dbSlaveDC2Mock.ExpectQuery(`.*pg_is_in_recovery.*`).WillReturnRows( | ||||
| 		sqlmock.NewRowsWithColumnDefinition( | ||||
| 			sqlmock.NewColumn("role").OfType("int8", 0), | ||||
| 			sqlmock.NewColumn("replication_lag").OfType("int8", 0)). | ||||
| 			AddRow(2, 50)). | ||||
| 		RowsWillBeClosed(). | ||||
| 		WithoutArgs() | ||||
|  | ||||
| 	dbSlaveDC2Mock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("slave-dc1")) | ||||
|  | ||||
| 	tctx, cancel := context.WithTimeout(t.Context(), 10*time.Second) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	c, err := NewCluster[Querier]( | ||||
| 		WithClusterContext(tctx), | ||||
| 		WithClusterNodeChecker(hasql.PostgreSQLChecker), | ||||
| 		WithClusterNodePicker(NewCustomPicker[Querier]( | ||||
| 			CustomPickerMaxLag(100), | ||||
| 		)), | ||||
| 		WithClusterNodes( | ||||
| 			ClusterNode{"slave-dc1", dbSlaveDC1, 1}, | ||||
| 			ClusterNode{"master-dc1", dbMaster, 1}, | ||||
| 			ClusterNode{"slave-dc2", dbSlaveDC2, 2}, | ||||
| 			ClusterNode{"drmaster1-dc2", dbDRMaster, 0}, | ||||
| 		), | ||||
| 		WithClusterOptions( | ||||
| 			hasql.WithUpdateInterval[Querier](2*time.Second), | ||||
| 			hasql.WithUpdateTimeout[Querier](1*time.Second), | ||||
| 		), | ||||
| 	) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer c.Close() | ||||
|  | ||||
| 	if err = c.WaitForNodes(tctx, hasql.Primary, hasql.Standby); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	time.Sleep(500 * time.Millisecond) | ||||
|  | ||||
| 	node1Name := "" | ||||
| 	fmt.Printf("check for Standby\n") | ||||
| 	if row := c.QueryRowContext(NodeStateCriterion(tctx, hasql.Standby), "SELECT node_name as name"); row.Err() != nil { | ||||
| 		t.Fatal(row.Err()) | ||||
| 	} else if err = row.Scan(&node1Name); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} else if "slave-dc1" != node1Name { | ||||
| 		t.Fatalf("invalid node name %s != %s", "slave-dc1", node1Name) | ||||
| 	} | ||||
|  | ||||
| 	dbSlaveDC1Mock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("slave-dc1")) | ||||
|  | ||||
| 	node2Name := "" | ||||
| 	fmt.Printf("check for PreferStandby\n") | ||||
| 	if row := c.QueryRowContext(NodeStateCriterion(tctx, hasql.PreferStandby), "SELECT node_name as name"); row.Err() != nil { | ||||
| 		t.Fatal(row.Err()) | ||||
| 	} else if err = row.Scan(&node2Name); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} else if "slave-dc1" != node2Name { | ||||
| 		t.Fatalf("invalid node name %s != %s", "slave-dc1", node2Name) | ||||
| 	} | ||||
|  | ||||
| 	node3Name := "" | ||||
| 	fmt.Printf("check for PreferPrimary\n") | ||||
| 	if row := c.QueryRowContext(NodeStateCriterion(tctx, hasql.PreferPrimary), "SELECT node_name as name"); row.Err() != nil { | ||||
| 		t.Fatal(row.Err()) | ||||
| 	} else if err = row.Scan(&node3Name); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} else if "master-dc1" != node3Name { | ||||
| 		t.Fatalf("invalid node name %s != %s", "master-dc1", node3Name) | ||||
| 	} | ||||
|  | ||||
| 	dbSlaveDC1Mock.ExpectQuery(`.*`).WillReturnRows(sqlmock.NewRows([]string{"role"}).RowError(1, fmt.Errorf("row error"))) | ||||
|  | ||||
| 	time.Sleep(2 * time.Second) | ||||
|  | ||||
| 	fmt.Printf("check for PreferStandby\n") | ||||
| 	if row := c.QueryRowContext(NodeStateCriterion(tctx, hasql.PreferStandby), "SELECT node_name as name"); row.Err() == nil { | ||||
| 		t.Fatal("must return error") | ||||
| 	} | ||||
|  | ||||
| 	if dbMasterErr := dbMasterMock.ExpectationsWereMet(); dbMasterErr != nil { | ||||
| 		t.Error(dbMasterErr) | ||||
| 	} | ||||
| } | ||||
							
								
								
									
										25
									
								
								cluster/hasql/db.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										25
									
								
								cluster/hasql/db.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,25 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"database/sql" | ||||
| ) | ||||
|  | ||||
| type Querier interface { | ||||
| 	// Basic connection methods | ||||
| 	PingContext(ctx context.Context) error | ||||
| 	Close() error | ||||
|  | ||||
| 	// Query methods with context | ||||
| 	ExecContext(ctx context.Context, query string, args ...interface{}) (sql.Result, error) | ||||
| 	QueryContext(ctx context.Context, query string, args ...interface{}) (*sql.Rows, error) | ||||
| 	QueryRowContext(ctx context.Context, query string, args ...interface{}) *sql.Row | ||||
|  | ||||
| 	// Prepared statements with context | ||||
| 	PrepareContext(ctx context.Context, query string) (*sql.Stmt, error) | ||||
|  | ||||
| 	// Transaction management with context | ||||
| 	BeginTx(ctx context.Context, opts *sql.TxOptions) (*sql.Tx, error) | ||||
|  | ||||
| 	Conn(ctx context.Context) (*sql.Conn, error) | ||||
| } | ||||
							
								
								
									
										295
									
								
								cluster/hasql/driver.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										295
									
								
								cluster/hasql/driver.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,295 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"database/sql" | ||||
| 	"database/sql/driver" | ||||
| 	"io" | ||||
| 	"sync" | ||||
| 	"time" | ||||
| ) | ||||
|  | ||||
| // OpenDBWithCluster creates a [*sql.DB] that uses the [ClusterQuerier] | ||||
| func OpenDBWithCluster(db ClusterQuerier) (*sql.DB, error) { | ||||
| 	driver := NewClusterDriver(db) | ||||
| 	connector, err := driver.OpenConnector("") | ||||
| 	if err != nil { | ||||
| 		return nil, err | ||||
| 	} | ||||
| 	return sql.OpenDB(connector), nil | ||||
| } | ||||
|  | ||||
| // ClusterDriver implements [driver.Driver] and driver.Connector for an existing [Querier] | ||||
| type ClusterDriver struct { | ||||
| 	db ClusterQuerier | ||||
| } | ||||
|  | ||||
| // NewClusterDriver creates a new [driver.Driver] that uses an existing [ClusterQuerier] | ||||
| func NewClusterDriver(db ClusterQuerier) *ClusterDriver { | ||||
| 	return &ClusterDriver{db: db} | ||||
| } | ||||
|  | ||||
| // Open implements [driver.Driver.Open] | ||||
| func (d *ClusterDriver) Open(name string) (driver.Conn, error) { | ||||
| 	return d.Connect(context.Background()) | ||||
| } | ||||
|  | ||||
| // OpenConnector implements [driver.DriverContext.OpenConnector] | ||||
| func (d *ClusterDriver) OpenConnector(name string) (driver.Connector, error) { | ||||
| 	return d, nil | ||||
| } | ||||
|  | ||||
| // Connect implements [driver.Connector.Connect] | ||||
| func (d *ClusterDriver) Connect(ctx context.Context) (driver.Conn, error) { | ||||
| 	conn, err := d.db.Conn(ctx) | ||||
| 	if err != nil { | ||||
| 		return nil, err | ||||
| 	} | ||||
| 	return &dbConn{conn: conn}, nil | ||||
| } | ||||
|  | ||||
| // Driver implements [driver.Connector.Driver] | ||||
| func (d *ClusterDriver) Driver() driver.Driver { | ||||
| 	return d | ||||
| } | ||||
|  | ||||
| // dbConn implements driver.Conn with both context and legacy methods | ||||
| type dbConn struct { | ||||
| 	conn *sql.Conn | ||||
| 	mu   sync.Mutex | ||||
| } | ||||
|  | ||||
| // Prepare implements [driver.Conn.Prepare] (legacy method) | ||||
| func (c *dbConn) Prepare(query string) (driver.Stmt, error) { | ||||
| 	return c.PrepareContext(context.Background(), query) | ||||
| } | ||||
|  | ||||
| // PrepareContext implements [driver.ConnPrepareContext.PrepareContext] | ||||
| func (c *dbConn) PrepareContext(ctx context.Context, query string) (driver.Stmt, error) { | ||||
| 	c.mu.Lock() | ||||
| 	defer c.mu.Unlock() | ||||
|  | ||||
| 	stmt, err := c.conn.PrepareContext(ctx, query) | ||||
| 	if err != nil { | ||||
| 		return nil, err | ||||
| 	} | ||||
|  | ||||
| 	return &dbStmt{stmt: stmt}, nil | ||||
| } | ||||
|  | ||||
| // Exec implements [driver.Execer.Exec] (legacy method) | ||||
| func (c *dbConn) Exec(query string, args []driver.Value) (driver.Result, error) { | ||||
| 	namedArgs := make([]driver.NamedValue, len(args)) | ||||
| 	for i, value := range args { | ||||
| 		namedArgs[i] = driver.NamedValue{Value: value} | ||||
| 	} | ||||
| 	return c.ExecContext(context.Background(), query, namedArgs) | ||||
| } | ||||
|  | ||||
| // ExecContext implements [driver.ExecerContext.ExecContext] | ||||
| func (c *dbConn) ExecContext(ctx context.Context, query string, args []driver.NamedValue) (driver.Result, error) { | ||||
| 	c.mu.Lock() | ||||
| 	defer c.mu.Unlock() | ||||
|  | ||||
| 	// Convert driver.NamedValue to any | ||||
| 	interfaceArgs := make([]any, len(args)) | ||||
| 	for i, arg := range args { | ||||
| 		interfaceArgs[i] = arg.Value | ||||
| 	} | ||||
|  | ||||
| 	return c.conn.ExecContext(ctx, query, interfaceArgs...) | ||||
| } | ||||
|  | ||||
| // Query implements [driver.Queryer.Query] (legacy method) | ||||
| func (c *dbConn) Query(query string, args []driver.Value) (driver.Rows, error) { | ||||
| 	namedArgs := make([]driver.NamedValue, len(args)) | ||||
| 	for i, value := range args { | ||||
| 		namedArgs[i] = driver.NamedValue{Value: value} | ||||
| 	} | ||||
| 	return c.QueryContext(context.Background(), query, namedArgs) | ||||
| } | ||||
|  | ||||
| // QueryContext implements [driver.QueryerContext.QueryContext] | ||||
| func (c *dbConn) QueryContext(ctx context.Context, query string, args []driver.NamedValue) (driver.Rows, error) { | ||||
| 	c.mu.Lock() | ||||
| 	defer c.mu.Unlock() | ||||
|  | ||||
| 	// Convert driver.NamedValue to any | ||||
| 	interfaceArgs := make([]any, len(args)) | ||||
| 	for i, arg := range args { | ||||
| 		interfaceArgs[i] = arg.Value | ||||
| 	} | ||||
|  | ||||
| 	rows, err := c.conn.QueryContext(ctx, query, interfaceArgs...) | ||||
| 	if err != nil { | ||||
| 		return nil, err | ||||
| 	} | ||||
|  | ||||
| 	return &dbRows{rows: rows}, nil | ||||
| } | ||||
|  | ||||
| // Begin implements [driver.Conn.Begin] (legacy method) | ||||
| func (c *dbConn) Begin() (driver.Tx, error) { | ||||
| 	return c.BeginTx(context.Background(), driver.TxOptions{}) | ||||
| } | ||||
|  | ||||
| // BeginTx implements [driver.ConnBeginTx.BeginTx] | ||||
| func (c *dbConn) BeginTx(ctx context.Context, opts driver.TxOptions) (driver.Tx, error) { | ||||
| 	c.mu.Lock() | ||||
| 	defer c.mu.Unlock() | ||||
|  | ||||
| 	sqlOpts := &sql.TxOptions{ | ||||
| 		Isolation: sql.IsolationLevel(opts.Isolation), | ||||
| 		ReadOnly:  opts.ReadOnly, | ||||
| 	} | ||||
|  | ||||
| 	tx, err := c.conn.BeginTx(ctx, sqlOpts) | ||||
| 	if err != nil { | ||||
| 		return nil, err | ||||
| 	} | ||||
|  | ||||
| 	return &dbTx{tx: tx}, nil | ||||
| } | ||||
|  | ||||
| // Ping implements [driver.Pinger.Ping] | ||||
| func (c *dbConn) Ping(ctx context.Context) error { | ||||
| 	return c.conn.PingContext(ctx) | ||||
| } | ||||
|  | ||||
| // Close implements [driver.Conn.Close] | ||||
| func (c *dbConn) Close() error { | ||||
| 	return c.conn.Close() | ||||
| } | ||||
|  | ||||
| // IsValid implements [driver.Validator.IsValid] | ||||
| func (c *dbConn) IsValid() bool { | ||||
| 	// Ping with a short timeout to check if the connection is still valid | ||||
| 	ctx, cancel := context.WithTimeout(context.Background(), 1*time.Second) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	return c.conn.PingContext(ctx) == nil | ||||
| } | ||||
|  | ||||
| // dbStmt implements [driver.Stmt] with both context and legacy methods | ||||
| type dbStmt struct { | ||||
| 	stmt *sql.Stmt | ||||
| 	mu   sync.Mutex | ||||
| } | ||||
|  | ||||
| // Close implements [driver.Stmt.Close] | ||||
| func (s *dbStmt) Close() error { | ||||
| 	s.mu.Lock() | ||||
| 	defer s.mu.Unlock() | ||||
| 	return s.stmt.Close() | ||||
| } | ||||
|  | ||||
| // Close implements [driver.Stmt.NumInput] | ||||
| func (s *dbStmt) NumInput() int { | ||||
| 	return -1 // Number of parameters is unknown | ||||
| } | ||||
|  | ||||
| // Exec implements [driver.Stmt.Exec] (legacy method) | ||||
| func (s *dbStmt) Exec(args []driver.Value) (driver.Result, error) { | ||||
| 	namedArgs := make([]driver.NamedValue, len(args)) | ||||
| 	for i, value := range args { | ||||
| 		namedArgs[i] = driver.NamedValue{Value: value} | ||||
| 	} | ||||
| 	return s.ExecContext(context.Background(), namedArgs) | ||||
| } | ||||
|  | ||||
| // ExecContext implements [driver.StmtExecContext.ExecContext] | ||||
| func (s *dbStmt) ExecContext(ctx context.Context, args []driver.NamedValue) (driver.Result, error) { | ||||
| 	s.mu.Lock() | ||||
| 	defer s.mu.Unlock() | ||||
|  | ||||
| 	interfaceArgs := make([]any, len(args)) | ||||
| 	for i, arg := range args { | ||||
| 		interfaceArgs[i] = arg.Value | ||||
| 	} | ||||
| 	return s.stmt.ExecContext(ctx, interfaceArgs...) | ||||
| } | ||||
|  | ||||
| // Query implements [driver.Stmt.Query] (legacy method) | ||||
| func (s *dbStmt) Query(args []driver.Value) (driver.Rows, error) { | ||||
| 	namedArgs := make([]driver.NamedValue, len(args)) | ||||
| 	for i, value := range args { | ||||
| 		namedArgs[i] = driver.NamedValue{Value: value} | ||||
| 	} | ||||
| 	return s.QueryContext(context.Background(), namedArgs) | ||||
| } | ||||
|  | ||||
| // QueryContext implements [driver.StmtQueryContext.QueryContext] | ||||
| func (s *dbStmt) QueryContext(ctx context.Context, args []driver.NamedValue) (driver.Rows, error) { | ||||
| 	s.mu.Lock() | ||||
| 	defer s.mu.Unlock() | ||||
|  | ||||
| 	interfaceArgs := make([]any, len(args)) | ||||
| 	for i, arg := range args { | ||||
| 		interfaceArgs[i] = arg.Value | ||||
| 	} | ||||
|  | ||||
| 	rows, err := s.stmt.QueryContext(ctx, interfaceArgs...) | ||||
| 	if err != nil { | ||||
| 		return nil, err | ||||
| 	} | ||||
|  | ||||
| 	return &dbRows{rows: rows}, nil | ||||
| } | ||||
|  | ||||
| // dbRows implements [driver.Rows] | ||||
| type dbRows struct { | ||||
| 	rows *sql.Rows | ||||
| } | ||||
|  | ||||
| // Columns implements [driver.Rows.Columns] | ||||
| func (r *dbRows) Columns() []string { | ||||
| 	cols, err := r.rows.Columns() | ||||
| 	if err != nil { | ||||
| 		// This shouldn't happen if the query was successful | ||||
| 		return []string{} | ||||
| 	} | ||||
| 	return cols | ||||
| } | ||||
|  | ||||
| // Close implements [driver.Rows.Close] | ||||
| func (r *dbRows) Close() error { | ||||
| 	return r.rows.Close() | ||||
| } | ||||
|  | ||||
| // Next implements [driver.Rows.Next] | ||||
| func (r *dbRows) Next(dest []driver.Value) error { | ||||
| 	if !r.rows.Next() { | ||||
| 		if err := r.rows.Err(); err != nil { | ||||
| 			return err | ||||
| 		} | ||||
| 		return io.EOF | ||||
| 	} | ||||
|  | ||||
| 	// Create a slice of interfaces to scan into | ||||
| 	scanArgs := make([]any, len(dest)) | ||||
| 	for i := range scanArgs { | ||||
| 		scanArgs[i] = &dest[i] | ||||
| 	} | ||||
|  | ||||
| 	return r.rows.Scan(scanArgs...) | ||||
| } | ||||
|  | ||||
| // dbTx implements [driver.Tx] | ||||
| type dbTx struct { | ||||
| 	tx *sql.Tx | ||||
| 	mu sync.Mutex | ||||
| } | ||||
|  | ||||
| // Commit implements [driver.Tx.Commit] | ||||
| func (t *dbTx) Commit() error { | ||||
| 	t.mu.Lock() | ||||
| 	defer t.mu.Unlock() | ||||
| 	return t.tx.Commit() | ||||
| } | ||||
|  | ||||
| // Rollback implements [driver.Tx.Rollback] | ||||
| func (t *dbTx) Rollback() error { | ||||
| 	t.mu.Lock() | ||||
| 	defer t.mu.Unlock() | ||||
| 	return t.tx.Rollback() | ||||
| } | ||||
							
								
								
									
										141
									
								
								cluster/hasql/driver_test.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										141
									
								
								cluster/hasql/driver_test.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,141 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"testing" | ||||
| 	"time" | ||||
|  | ||||
| 	"github.com/DATA-DOG/go-sqlmock" | ||||
| 	"golang.yandex/hasql/v2" | ||||
| ) | ||||
|  | ||||
| func TestDriver(t *testing.T) { | ||||
| 	dbMaster, dbMasterMock, err := sqlmock.New(sqlmock.MonitorPingsOption(true)) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer dbMaster.Close() | ||||
| 	dbMasterMock.MatchExpectationsInOrder(false) | ||||
|  | ||||
| 	dbMasterMock.ExpectQuery(`.*pg_is_in_recovery.*`).WillReturnRows( | ||||
| 		sqlmock.NewRowsWithColumnDefinition( | ||||
| 			sqlmock.NewColumn("role").OfType("int8", 0), | ||||
| 			sqlmock.NewColumn("replication_lag").OfType("int8", 0)). | ||||
| 			AddRow(1, 0)). | ||||
| 		RowsWillBeClosed(). | ||||
| 		WithoutArgs() | ||||
|  | ||||
| 	dbMasterMock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("master-dc1")) | ||||
|  | ||||
| 	dbDRMaster, dbDRMasterMock, err := sqlmock.New(sqlmock.MonitorPingsOption(true)) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer dbDRMaster.Close() | ||||
| 	dbDRMasterMock.MatchExpectationsInOrder(false) | ||||
|  | ||||
| 	dbDRMasterMock.ExpectQuery(`.*pg_is_in_recovery.*`).WillReturnRows( | ||||
| 		sqlmock.NewRowsWithColumnDefinition( | ||||
| 			sqlmock.NewColumn("role").OfType("int8", 0), | ||||
| 			sqlmock.NewColumn("replication_lag").OfType("int8", 0)). | ||||
| 			AddRow(2, 40)). | ||||
| 		RowsWillBeClosed(). | ||||
| 		WithoutArgs() | ||||
|  | ||||
| 	dbDRMasterMock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("drmaster1-dc2")) | ||||
|  | ||||
| 	dbDRMasterMock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("drmaster")) | ||||
|  | ||||
| 	dbSlaveDC1, dbSlaveDC1Mock, err := sqlmock.New(sqlmock.MonitorPingsOption(true)) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer dbSlaveDC1.Close() | ||||
| 	dbSlaveDC1Mock.MatchExpectationsInOrder(false) | ||||
|  | ||||
| 	dbSlaveDC1Mock.ExpectQuery(`.*pg_is_in_recovery.*`).WillReturnRows( | ||||
| 		sqlmock.NewRowsWithColumnDefinition( | ||||
| 			sqlmock.NewColumn("role").OfType("int8", 0), | ||||
| 			sqlmock.NewColumn("replication_lag").OfType("int8", 0)). | ||||
| 			AddRow(2, 50)). | ||||
| 		RowsWillBeClosed(). | ||||
| 		WithoutArgs() | ||||
|  | ||||
| 	dbSlaveDC1Mock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("slave-dc1")) | ||||
|  | ||||
| 	dbSlaveDC2, dbSlaveDC2Mock, err := sqlmock.New(sqlmock.MonitorPingsOption(true)) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer dbSlaveDC2.Close() | ||||
| 	dbSlaveDC1Mock.MatchExpectationsInOrder(false) | ||||
|  | ||||
| 	dbSlaveDC2Mock.ExpectQuery(`.*pg_is_in_recovery.*`).WillReturnRows( | ||||
| 		sqlmock.NewRowsWithColumnDefinition( | ||||
| 			sqlmock.NewColumn("role").OfType("int8", 0), | ||||
| 			sqlmock.NewColumn("replication_lag").OfType("int8", 0)). | ||||
| 			AddRow(2, 50)). | ||||
| 		RowsWillBeClosed(). | ||||
| 		WithoutArgs() | ||||
|  | ||||
| 	dbSlaveDC2Mock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("slave-dc1")) | ||||
|  | ||||
| 	tctx, cancel := context.WithTimeout(t.Context(), 10*time.Second) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	c, err := NewCluster[Querier]( | ||||
| 		WithClusterContext(tctx), | ||||
| 		WithClusterNodeChecker(hasql.PostgreSQLChecker), | ||||
| 		WithClusterNodePicker(NewCustomPicker[Querier]( | ||||
| 			CustomPickerMaxLag(100), | ||||
| 		)), | ||||
| 		WithClusterNodes( | ||||
| 			ClusterNode{"slave-dc1", dbSlaveDC1, 1}, | ||||
| 			ClusterNode{"master-dc1", dbMaster, 1}, | ||||
| 			ClusterNode{"slave-dc2", dbSlaveDC2, 2}, | ||||
| 			ClusterNode{"drmaster1-dc2", dbDRMaster, 0}, | ||||
| 		), | ||||
| 		WithClusterOptions( | ||||
| 			hasql.WithUpdateInterval[Querier](2*time.Second), | ||||
| 			hasql.WithUpdateTimeout[Querier](1*time.Second), | ||||
| 		), | ||||
| 	) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer c.Close() | ||||
|  | ||||
| 	if err = c.WaitForNodes(tctx, hasql.Primary, hasql.Standby); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	db, err := OpenDBWithCluster(c) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	// Use context methods | ||||
| 	row := db.QueryRowContext(NodeStateCriterion(t.Context(), hasql.Primary), "SELECT node_name as name") | ||||
| 	if err = row.Err(); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	nodeName := "" | ||||
| 	if err = row.Scan(&nodeName); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	if nodeName != "master-dc1" { | ||||
| 		t.Fatalf("invalid node_name %s != %s", "master-dc1", nodeName) | ||||
| 	} | ||||
| } | ||||
							
								
								
									
										10
									
								
								cluster/hasql/error.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										10
									
								
								cluster/hasql/error.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,10 @@ | ||||
| package sql | ||||
|  | ||||
| import "errors" | ||||
|  | ||||
| var ( | ||||
| 	ErrClusterChecker    = errors.New("cluster node checker required") | ||||
| 	ErrClusterDiscoverer = errors.New("cluster node discoverer required") | ||||
| 	ErrClusterPicker     = errors.New("cluster node picker required") | ||||
| 	ErrorNoAliveNodes    = errors.New("cluster no alive nodes") | ||||
| ) | ||||
							
								
								
									
										110
									
								
								cluster/hasql/options.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										110
									
								
								cluster/hasql/options.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,110 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"math" | ||||
|  | ||||
| 	"golang.yandex/hasql/v2" | ||||
| ) | ||||
|  | ||||
| // ClusterOptions contains cluster specific options | ||||
| type ClusterOptions struct { | ||||
| 	NodeChecker        hasql.NodeChecker | ||||
| 	NodePicker         hasql.NodePicker[Querier] | ||||
| 	NodeDiscoverer     hasql.NodeDiscoverer[Querier] | ||||
| 	Options            []hasql.ClusterOpt[Querier] | ||||
| 	Context            context.Context | ||||
| 	Retries            int | ||||
| 	NodePriority       map[string]int32 | ||||
| 	NodeStateCriterion hasql.NodeStateCriterion | ||||
| } | ||||
|  | ||||
| // ClusterOption apply cluster options to ClusterOptions | ||||
| type ClusterOption func(*ClusterOptions) | ||||
|  | ||||
| // WithClusterNodeChecker pass hasql.NodeChecker to cluster options | ||||
| func WithClusterNodeChecker(c hasql.NodeChecker) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.NodeChecker = c | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClusterNodePicker pass hasql.NodePicker to cluster options | ||||
| func WithClusterNodePicker(p hasql.NodePicker[Querier]) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.NodePicker = p | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClusterNodeDiscoverer pass hasql.NodeDiscoverer to cluster options | ||||
| func WithClusterNodeDiscoverer(d hasql.NodeDiscoverer[Querier]) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.NodeDiscoverer = d | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithRetries retry count on other nodes in case of error | ||||
| func WithRetries(n int) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.Retries = n | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClusterContext pass context.Context to cluster options and used for checks | ||||
| func WithClusterContext(ctx context.Context) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.Context = ctx | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClusterOptions pass hasql.ClusterOpt | ||||
| func WithClusterOptions(opts ...hasql.ClusterOpt[Querier]) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.Options = append(o.Options, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClusterNodeStateCriterion pass default hasql.NodeStateCriterion | ||||
| func WithClusterNodeStateCriterion(c hasql.NodeStateCriterion) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.NodeStateCriterion = c | ||||
| 	} | ||||
| } | ||||
|  | ||||
| type ClusterNode struct { | ||||
| 	Name     string | ||||
| 	DB       Querier | ||||
| 	Priority int32 | ||||
| } | ||||
|  | ||||
| // WithClusterNodes create cluster with static NodeDiscoverer | ||||
| func WithClusterNodes(cns ...ClusterNode) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		nodes := make([]*hasql.Node[Querier], 0, len(cns)) | ||||
| 		if o.NodePriority == nil { | ||||
| 			o.NodePriority = make(map[string]int32, len(cns)) | ||||
| 		} | ||||
| 		for _, cn := range cns { | ||||
| 			nodes = append(nodes, hasql.NewNode(cn.Name, cn.DB)) | ||||
| 			if cn.Priority == 0 { | ||||
| 				cn.Priority = math.MaxInt32 | ||||
| 			} | ||||
| 			o.NodePriority[cn.Name] = cn.Priority | ||||
| 		} | ||||
| 		o.NodeDiscoverer = hasql.NewStaticNodeDiscoverer(nodes...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| type nodeStateCriterionKey struct{} | ||||
|  | ||||
| // NodeStateCriterion inject hasql.NodeStateCriterion to context | ||||
| func NodeStateCriterion(ctx context.Context, c hasql.NodeStateCriterion) context.Context { | ||||
| 	return context.WithValue(ctx, nodeStateCriterionKey{}, c) | ||||
| } | ||||
|  | ||||
| func (c *Cluster) getNodeStateCriterion(ctx context.Context) hasql.NodeStateCriterion { | ||||
| 	if v, ok := ctx.Value(nodeStateCriterionKey{}).(hasql.NodeStateCriterion); ok { | ||||
| 		return v | ||||
| 	} | ||||
| 	return c.options.NodeStateCriterion | ||||
| } | ||||
							
								
								
									
										113
									
								
								cluster/hasql/picker.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										113
									
								
								cluster/hasql/picker.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,113 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"fmt" | ||||
| 	"math" | ||||
| 	"time" | ||||
|  | ||||
| 	"golang.yandex/hasql/v2" | ||||
| ) | ||||
|  | ||||
| // compile time guard | ||||
| var _ hasql.NodePicker[Querier] = (*CustomPicker[Querier])(nil) | ||||
|  | ||||
| // CustomPickerOptions holds options to pick nodes | ||||
| type CustomPickerOptions struct { | ||||
| 	MaxLag   int | ||||
| 	Priority map[string]int32 | ||||
| 	Retries  int | ||||
| } | ||||
|  | ||||
| // CustomPickerOption func apply option to CustomPickerOptions | ||||
| type CustomPickerOption func(*CustomPickerOptions) | ||||
|  | ||||
| // CustomPickerMaxLag specifies max lag for which node can be used | ||||
| func CustomPickerMaxLag(n int) CustomPickerOption { | ||||
| 	return func(o *CustomPickerOptions) { | ||||
| 		o.MaxLag = n | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // NewCustomPicker creates new node picker | ||||
| func NewCustomPicker[T Querier](opts ...CustomPickerOption) *CustomPicker[Querier] { | ||||
| 	options := CustomPickerOptions{} | ||||
| 	for _, o := range opts { | ||||
| 		o(&options) | ||||
| 	} | ||||
| 	return &CustomPicker[Querier]{opts: options} | ||||
| } | ||||
|  | ||||
| // CustomPicker holds node picker options | ||||
| type CustomPicker[T Querier] struct { | ||||
| 	opts CustomPickerOptions | ||||
| } | ||||
|  | ||||
| // PickNode used to return specific node | ||||
| func (p *CustomPicker[T]) PickNode(cnodes []hasql.CheckedNode[T]) hasql.CheckedNode[T] { | ||||
| 	for _, n := range cnodes { | ||||
| 		fmt.Printf("node %s\n", n.Node.String()) | ||||
| 	} | ||||
| 	return cnodes[0] | ||||
| } | ||||
|  | ||||
| func (p *CustomPicker[T]) getPriority(nodeName string) int32 { | ||||
| 	if prio, ok := p.opts.Priority[nodeName]; ok { | ||||
| 		return prio | ||||
| 	} | ||||
| 	return math.MaxInt32 // Default to lowest priority | ||||
| } | ||||
|  | ||||
| // CompareNodes used to sort nodes | ||||
| func (p *CustomPicker[T]) CompareNodes(a, b hasql.CheckedNode[T]) int { | ||||
| 	// Get replication lag values | ||||
| 	aLag := a.Info.(interface{ ReplicationLag() int }).ReplicationLag() | ||||
| 	bLag := b.Info.(interface{ ReplicationLag() int }).ReplicationLag() | ||||
|  | ||||
| 	// First check that lag lower then MaxLag | ||||
| 	if aLag > p.opts.MaxLag && bLag > p.opts.MaxLag { | ||||
| 		return 0 // both are equal | ||||
| 	} | ||||
|  | ||||
| 	// If one node exceeds MaxLag and the other doesn't, prefer the one that doesn't | ||||
| 	if aLag > p.opts.MaxLag { | ||||
| 		return 1 // b is better | ||||
| 	} | ||||
| 	if bLag > p.opts.MaxLag { | ||||
| 		return -1 // a is better | ||||
| 	} | ||||
|  | ||||
| 	// Get node priorities | ||||
| 	aPrio := p.getPriority(a.Node.String()) | ||||
| 	bPrio := p.getPriority(b.Node.String()) | ||||
|  | ||||
| 	// if both priority equals | ||||
| 	if aPrio == bPrio { | ||||
| 		// First compare by replication lag | ||||
| 		if aLag < bLag { | ||||
| 			return -1 | ||||
| 		} | ||||
| 		if aLag > bLag { | ||||
| 			return 1 | ||||
| 		} | ||||
| 		// If replication lag is equal, compare by latency | ||||
| 		aLatency := a.Info.(interface{ Latency() time.Duration }).Latency() | ||||
| 		bLatency := b.Info.(interface{ Latency() time.Duration }).Latency() | ||||
|  | ||||
| 		if aLatency < bLatency { | ||||
| 			return -1 | ||||
| 		} | ||||
| 		if aLatency > bLatency { | ||||
| 			return 1 | ||||
| 		} | ||||
|  | ||||
| 		// If lag and latency is equal | ||||
| 		return 0 | ||||
| 	} | ||||
|  | ||||
| 	// If priorities are different, prefer the node with lower priority value | ||||
| 	if aPrio < bPrio { | ||||
| 		return -1 | ||||
| 	} | ||||
|  | ||||
| 	return 1 | ||||
| } | ||||
							
								
								
									
										531
									
								
								cluster/sql/cluster.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										531
									
								
								cluster/sql/cluster.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,531 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"database/sql" | ||||
| 	"errors" | ||||
| 	"fmt" | ||||
| 	"math" | ||||
| 	"reflect" | ||||
| 	"time" | ||||
| 	"unsafe" | ||||
|  | ||||
| 	"golang.yandex/hasql/v2" | ||||
| ) | ||||
|  | ||||
| var errNoAliveNodes = errors.New("no alive nodes") | ||||
|  | ||||
| func newSQLRowError() *sql.Row { | ||||
| 	row := &sql.Row{} | ||||
| 	t := reflect.TypeOf(row).Elem() | ||||
| 	field, _ := t.FieldByName("err") | ||||
| 	rowPtr := unsafe.Pointer(row) | ||||
| 	errFieldPtr := unsafe.Pointer(uintptr(rowPtr) + field.Offset) | ||||
| 	errPtr := (*error)(errFieldPtr) | ||||
| 	*errPtr = errNoAliveNodes | ||||
| 	return row | ||||
| } | ||||
|  | ||||
| type ClusterQuerier interface { | ||||
| 	Querier | ||||
| 	WaitForNodes(ctx context.Context, criterion ...hasql.NodeStateCriterion) error | ||||
| } | ||||
|  | ||||
| type Querier interface { | ||||
| 	// Basic connection methods | ||||
| 	PingContext(ctx context.Context) error | ||||
| 	Close() error | ||||
|  | ||||
| 	// Query methods with context | ||||
| 	ExecContext(ctx context.Context, query string, args ...interface{}) (sql.Result, error) | ||||
| 	QueryContext(ctx context.Context, query string, args ...interface{}) (*sql.Rows, error) | ||||
| 	QueryRowContext(ctx context.Context, query string, args ...interface{}) *sql.Row | ||||
|  | ||||
| 	// Prepared statements with context | ||||
| 	PrepareContext(ctx context.Context, query string) (*sql.Stmt, error) | ||||
|  | ||||
| 	// Transaction management with context | ||||
| 	BeginTx(ctx context.Context, opts *sql.TxOptions) (*sql.Tx, error) | ||||
|  | ||||
| 	// Connection pool management | ||||
| 	SetConnMaxLifetime(d time.Duration) | ||||
| 	SetConnMaxIdleTime(d time.Duration) | ||||
| 	SetMaxOpenConns(n int) | ||||
| 	SetMaxIdleConns(n int) | ||||
| 	Stats() sql.DBStats | ||||
|  | ||||
| 	Conn(ctx context.Context) (*sql.Conn, error) | ||||
| } | ||||
|  | ||||
| var ( | ||||
| 	ErrClusterChecker    = errors.New("cluster node checker required") | ||||
| 	ErrClusterDiscoverer = errors.New("cluster node discoverer required") | ||||
| 	ErrClusterPicker     = errors.New("cluster node picker required") | ||||
| ) | ||||
|  | ||||
| type Cluster struct { | ||||
| 	hasql   *hasql.Cluster[Querier] | ||||
| 	options ClusterOptions | ||||
| } | ||||
|  | ||||
| // NewCluster returns Querier that provides cluster of nodes | ||||
| func NewCluster[T Querier](opts ...ClusterOption) (ClusterQuerier, error) { | ||||
| 	options := ClusterOptions{Context: context.Background()} | ||||
| 	for _, opt := range opts { | ||||
| 		opt(&options) | ||||
| 	} | ||||
| 	if options.NodeChecker == nil { | ||||
| 		return nil, ErrClusterChecker | ||||
| 	} | ||||
| 	if options.NodeDiscoverer == nil { | ||||
| 		return nil, ErrClusterDiscoverer | ||||
| 	} | ||||
| 	if options.NodePicker == nil { | ||||
| 		return nil, ErrClusterPicker | ||||
| 	} | ||||
|  | ||||
| 	if options.Retries < 1 { | ||||
| 		options.Retries = 1 | ||||
| 	} | ||||
|  | ||||
| 	if options.NodeStateCriterion == 0 { | ||||
| 		options.NodeStateCriterion = hasql.Primary | ||||
| 	} | ||||
|  | ||||
| 	options.Options = append(options.Options, hasql.WithNodePicker(options.NodePicker)) | ||||
| 	if p, ok := options.NodePicker.(*CustomPicker[Querier]); ok { | ||||
| 		p.opts.Priority = options.NodePriority | ||||
| 	} | ||||
|  | ||||
| 	c, err := hasql.NewCluster( | ||||
| 		options.NodeDiscoverer, | ||||
| 		options.NodeChecker, | ||||
| 		options.Options..., | ||||
| 	) | ||||
| 	if err != nil { | ||||
| 		return nil, err | ||||
| 	} | ||||
|  | ||||
| 	return &Cluster{hasql: c, options: options}, nil | ||||
| } | ||||
|  | ||||
| // compile time guard | ||||
| var _ hasql.NodePicker[Querier] = (*CustomPicker[Querier])(nil) | ||||
|  | ||||
| type nodeStateCriterionKey struct{} | ||||
|  | ||||
| // NodeStateCriterion inject hasql.NodeStateCriterion to context | ||||
| func NodeStateCriterion(ctx context.Context, c hasql.NodeStateCriterion) context.Context { | ||||
| 	return context.WithValue(ctx, nodeStateCriterionKey{}, c) | ||||
| } | ||||
|  | ||||
| // CustomPickerOptions holds options to pick nodes | ||||
| type CustomPickerOptions struct { | ||||
| 	MaxLag   int | ||||
| 	Priority map[string]int32 | ||||
| 	Retries  int | ||||
| } | ||||
|  | ||||
| // CustomPickerOption func apply option to CustomPickerOptions | ||||
| type CustomPickerOption func(*CustomPickerOptions) | ||||
|  | ||||
| // CustomPickerMaxLag specifies max lag for which node can be used | ||||
| func CustomPickerMaxLag(n int) CustomPickerOption { | ||||
| 	return func(o *CustomPickerOptions) { | ||||
| 		o.MaxLag = n | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // NewCustomPicker creates new node picker | ||||
| func NewCustomPicker[T Querier](opts ...CustomPickerOption) *CustomPicker[Querier] { | ||||
| 	options := CustomPickerOptions{} | ||||
| 	for _, o := range opts { | ||||
| 		o(&options) | ||||
| 	} | ||||
| 	return &CustomPicker[Querier]{opts: options} | ||||
| } | ||||
|  | ||||
| // CustomPicker holds node picker options | ||||
| type CustomPicker[T Querier] struct { | ||||
| 	opts CustomPickerOptions | ||||
| } | ||||
|  | ||||
| // PickNode used to return specific node | ||||
| func (p *CustomPicker[T]) PickNode(cnodes []hasql.CheckedNode[T]) hasql.CheckedNode[T] { | ||||
| 	for _, n := range cnodes { | ||||
| 		fmt.Printf("node %s\n", n.Node.String()) | ||||
| 	} | ||||
| 	return cnodes[0] | ||||
| } | ||||
|  | ||||
| func (p *CustomPicker[T]) getPriority(nodeName string) int32 { | ||||
| 	if prio, ok := p.opts.Priority[nodeName]; ok { | ||||
| 		return prio | ||||
| 	} | ||||
| 	return math.MaxInt32 // Default to lowest priority | ||||
| } | ||||
|  | ||||
| // CompareNodes used to sort nodes | ||||
| func (p *CustomPicker[T]) CompareNodes(a, b hasql.CheckedNode[T]) int { | ||||
| 	fmt.Printf("CompareNodes %s %s\n", a.Node.String(), b.Node.String()) | ||||
| 	// Get replication lag values | ||||
| 	aLag := a.Info.(interface{ ReplicationLag() int }).ReplicationLag() | ||||
| 	bLag := b.Info.(interface{ ReplicationLag() int }).ReplicationLag() | ||||
|  | ||||
| 	// First check that lag lower then MaxLag | ||||
| 	if aLag > p.opts.MaxLag && bLag > p.opts.MaxLag { | ||||
| 		fmt.Printf("CompareNodes aLag > p.opts.MaxLag && bLag > p.opts.MaxLag\n") | ||||
| 		return 0 // both are equal | ||||
| 	} | ||||
|  | ||||
| 	// If one node exceeds MaxLag and the other doesn't, prefer the one that doesn't | ||||
| 	if aLag > p.opts.MaxLag { | ||||
| 		fmt.Printf("CompareNodes aLag > p.opts.MaxLag\n") | ||||
| 		return 1 // b is better | ||||
| 	} | ||||
| 	if bLag > p.opts.MaxLag { | ||||
| 		fmt.Printf("CompareNodes bLag > p.opts.MaxLag\n") | ||||
| 		return -1 // a is better | ||||
| 	} | ||||
|  | ||||
| 	// Get node priorities | ||||
| 	aPrio := p.getPriority(a.Node.String()) | ||||
| 	bPrio := p.getPriority(b.Node.String()) | ||||
|  | ||||
| 	// if both priority equals | ||||
| 	if aPrio == bPrio { | ||||
| 		fmt.Printf("CompareNodes aPrio == bPrio\n") | ||||
| 		// First compare by replication lag | ||||
| 		if aLag < bLag { | ||||
| 			fmt.Printf("CompareNodes aLag < bLag\n") | ||||
| 			return -1 | ||||
| 		} | ||||
| 		if aLag > bLag { | ||||
| 			fmt.Printf("CompareNodes aLag > bLag\n") | ||||
| 			return 1 | ||||
| 		} | ||||
| 		// If replication lag is equal, compare by latency | ||||
| 		aLatency := a.Info.(interface{ Latency() time.Duration }).Latency() | ||||
| 		bLatency := b.Info.(interface{ Latency() time.Duration }).Latency() | ||||
|  | ||||
| 		if aLatency < bLatency { | ||||
| 			return -1 | ||||
| 		} | ||||
| 		if aLatency > bLatency { | ||||
| 			return 1 | ||||
| 		} | ||||
|  | ||||
| 		// If lag and latency is equal | ||||
| 		return 0 | ||||
| 	} | ||||
|  | ||||
| 	// If priorities are different, prefer the node with lower priority value | ||||
| 	if aPrio < bPrio { | ||||
| 		return -1 | ||||
| 	} | ||||
|  | ||||
| 	return 1 | ||||
| } | ||||
|  | ||||
| // ClusterOptions contains cluster specific options | ||||
| type ClusterOptions struct { | ||||
| 	NodeChecker        hasql.NodeChecker | ||||
| 	NodePicker         hasql.NodePicker[Querier] | ||||
| 	NodeDiscoverer     hasql.NodeDiscoverer[Querier] | ||||
| 	Options            []hasql.ClusterOpt[Querier] | ||||
| 	Context            context.Context | ||||
| 	Retries            int | ||||
| 	NodePriority       map[string]int32 | ||||
| 	NodeStateCriterion hasql.NodeStateCriterion | ||||
| } | ||||
|  | ||||
| // ClusterOption apply cluster options to ClusterOptions | ||||
| type ClusterOption func(*ClusterOptions) | ||||
|  | ||||
| // WithClusterNodeChecker pass hasql.NodeChecker to cluster options | ||||
| func WithClusterNodeChecker(c hasql.NodeChecker) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.NodeChecker = c | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClusterNodePicker pass hasql.NodePicker to cluster options | ||||
| func WithClusterNodePicker(p hasql.NodePicker[Querier]) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.NodePicker = p | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClusterNodeDiscoverer pass hasql.NodeDiscoverer to cluster options | ||||
| func WithClusterNodeDiscoverer(d hasql.NodeDiscoverer[Querier]) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.NodeDiscoverer = d | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithRetries retry count on other nodes in case of error | ||||
| func WithRetries(n int) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.Retries = n | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClusterContext pass context.Context to cluster options and used for checks | ||||
| func WithClusterContext(ctx context.Context) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.Context = ctx | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClusterOptions pass hasql.ClusterOpt | ||||
| func WithClusterOptions(opts ...hasql.ClusterOpt[Querier]) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.Options = append(o.Options, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClusterNodeStateCriterion pass default hasql.NodeStateCriterion | ||||
| func WithClusterNodeStateCriterion(c hasql.NodeStateCriterion) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		o.NodeStateCriterion = c | ||||
| 	} | ||||
| } | ||||
|  | ||||
| type ClusterNode struct { | ||||
| 	Name     string | ||||
| 	DB       Querier | ||||
| 	Priority int32 | ||||
| } | ||||
|  | ||||
| // WithClusterNodes create cluster with static NodeDiscoverer | ||||
| func WithClusterNodes(cns ...ClusterNode) ClusterOption { | ||||
| 	return func(o *ClusterOptions) { | ||||
| 		nodes := make([]*hasql.Node[Querier], 0, len(cns)) | ||||
| 		if o.NodePriority == nil { | ||||
| 			o.NodePriority = make(map[string]int32, len(cns)) | ||||
| 		} | ||||
| 		for _, cn := range cns { | ||||
| 			nodes = append(nodes, hasql.NewNode(cn.Name, cn.DB)) | ||||
| 			if cn.Priority == 0 { | ||||
| 				cn.Priority = math.MaxInt32 | ||||
| 			} | ||||
| 			o.NodePriority[cn.Name] = cn.Priority | ||||
| 		} | ||||
| 		o.NodeDiscoverer = hasql.NewStaticNodeDiscoverer(nodes...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (c *Cluster) BeginTx(ctx context.Context, opts *sql.TxOptions) (*sql.Tx, error) { | ||||
| 	var tx *sql.Tx | ||||
| 	var err error | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			if tx, err = n.DB().BeginTx(ctx, opts); err != nil && retries >= c.options.Retries { | ||||
| 				return true | ||||
| 			} | ||||
| 		} | ||||
| 		return false | ||||
| 	}) | ||||
|  | ||||
| 	if tx == nil && err == nil { | ||||
| 		err = errNoAliveNodes | ||||
| 	} | ||||
|  | ||||
| 	return tx, err | ||||
| } | ||||
|  | ||||
| func (c *Cluster) Close() error { | ||||
| 	return c.hasql.Close() | ||||
| } | ||||
|  | ||||
| func (c *Cluster) Conn(ctx context.Context) (*sql.Conn, error) { | ||||
| 	var conn *sql.Conn | ||||
| 	var err error | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			if conn, err = n.DB().Conn(ctx); err != nil && retries >= c.options.Retries { | ||||
| 				return true | ||||
| 			} | ||||
| 		} | ||||
| 		return false | ||||
| 	}) | ||||
|  | ||||
| 	if conn == nil && err == nil { | ||||
| 		err = errNoAliveNodes | ||||
| 	} | ||||
|  | ||||
| 	return conn, err | ||||
| } | ||||
|  | ||||
| func (c *Cluster) ExecContext(ctx context.Context, query string, args ...interface{}) (sql.Result, error) { | ||||
| 	var res sql.Result | ||||
| 	var err error | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			if res, err = n.DB().ExecContext(ctx, query, args...); err != nil && retries >= c.options.Retries { | ||||
| 				return true | ||||
| 			} | ||||
| 		} | ||||
| 		return false | ||||
| 	}) | ||||
|  | ||||
| 	if res == nil && err == nil { | ||||
| 		err = errNoAliveNodes | ||||
| 	} | ||||
|  | ||||
| 	return res, err | ||||
| } | ||||
|  | ||||
| func (c *Cluster) PrepareContext(ctx context.Context, query string) (*sql.Stmt, error) { | ||||
| 	var res *sql.Stmt | ||||
| 	var err error | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			if res, err = n.DB().PrepareContext(ctx, query); err != nil && retries >= c.options.Retries { | ||||
| 				return true | ||||
| 			} | ||||
| 		} | ||||
| 		return false | ||||
| 	}) | ||||
|  | ||||
| 	if res == nil && err == nil { | ||||
| 		err = errNoAliveNodes | ||||
| 	} | ||||
|  | ||||
| 	return res, err | ||||
| } | ||||
|  | ||||
| func (c *Cluster) QueryContext(ctx context.Context, query string, args ...interface{}) (*sql.Rows, error) { | ||||
| 	var res *sql.Rows | ||||
| 	var err error | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			if res, err = n.DB().QueryContext(ctx, query); err != nil && err != sql.ErrNoRows && retries >= c.options.Retries { | ||||
| 				return true | ||||
| 			} | ||||
| 		} | ||||
| 		return false | ||||
| 	}) | ||||
|  | ||||
| 	if res == nil && err == nil { | ||||
| 		err = errNoAliveNodes | ||||
| 	} | ||||
|  | ||||
| 	return res, err | ||||
| } | ||||
|  | ||||
| func (c *Cluster) QueryRowContext(ctx context.Context, query string, args ...interface{}) *sql.Row { | ||||
| 	var res *sql.Row | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			res = n.DB().QueryRowContext(ctx, query, args...) | ||||
| 			if res.Err() == nil { | ||||
| 				return false | ||||
| 			} else if res.Err() != nil && retries >= c.options.Retries { | ||||
| 				return false | ||||
| 			} | ||||
| 		} | ||||
| 		return true | ||||
| 	}) | ||||
|  | ||||
| 	if res == nil { | ||||
| 		res = newSQLRowError() | ||||
| 	} | ||||
|  | ||||
| 	return res | ||||
| } | ||||
|  | ||||
| func (c *Cluster) PingContext(ctx context.Context) error { | ||||
| 	var err error | ||||
| 	var ok bool | ||||
|  | ||||
| 	retries := 0 | ||||
| 	c.hasql.NodesIter(c.getNodeStateCriterion(ctx))(func(n *hasql.Node[Querier]) bool { | ||||
| 		ok = true | ||||
| 		for ; retries < c.options.Retries; retries++ { | ||||
| 			if err = n.DB().PingContext(ctx); err != nil && retries >= c.options.Retries { | ||||
| 				return true | ||||
| 			} | ||||
| 		} | ||||
| 		return false | ||||
| 	}) | ||||
|  | ||||
| 	if !ok { | ||||
| 		err = errNoAliveNodes | ||||
| 	} | ||||
|  | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| func (c *Cluster) WaitForNodes(ctx context.Context, criterions ...hasql.NodeStateCriterion) error { | ||||
| 	for _, criterion := range criterions { | ||||
| 		if _, err := c.hasql.WaitForNode(ctx, criterion); err != nil { | ||||
| 			return err | ||||
| 		} | ||||
| 	} | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| func (c *Cluster) SetConnMaxLifetime(td time.Duration) { | ||||
| 	c.hasql.NodesIter(hasql.NodeStateCriterion(hasql.Alive))(func(n *hasql.Node[Querier]) bool { | ||||
| 		n.DB().SetConnMaxIdleTime(td) | ||||
| 		return false | ||||
| 	}) | ||||
| } | ||||
|  | ||||
| func (c *Cluster) SetConnMaxIdleTime(td time.Duration) { | ||||
| 	c.hasql.NodesIter(hasql.NodeStateCriterion(hasql.Alive))(func(n *hasql.Node[Querier]) bool { | ||||
| 		n.DB().SetConnMaxIdleTime(td) | ||||
| 		return false | ||||
| 	}) | ||||
| } | ||||
|  | ||||
| func (c *Cluster) SetMaxOpenConns(nc int) { | ||||
| 	c.hasql.NodesIter(hasql.NodeStateCriterion(hasql.Alive))(func(n *hasql.Node[Querier]) bool { | ||||
| 		n.DB().SetMaxOpenConns(nc) | ||||
| 		return false | ||||
| 	}) | ||||
| } | ||||
|  | ||||
| func (c *Cluster) SetMaxIdleConns(nc int) { | ||||
| 	c.hasql.NodesIter(hasql.NodeStateCriterion(hasql.Alive))(func(n *hasql.Node[Querier]) bool { | ||||
| 		n.DB().SetMaxIdleConns(nc) | ||||
| 		return false | ||||
| 	}) | ||||
| } | ||||
|  | ||||
| func (c *Cluster) Stats() sql.DBStats { | ||||
| 	s := sql.DBStats{} | ||||
| 	c.hasql.NodesIter(hasql.NodeStateCriterion(hasql.Alive))(func(n *hasql.Node[Querier]) bool { | ||||
| 		st := n.DB().Stats() | ||||
| 		s.Idle += st.Idle | ||||
| 		s.InUse += st.InUse | ||||
| 		s.MaxIdleClosed += st.MaxIdleClosed | ||||
| 		s.MaxIdleTimeClosed += st.MaxIdleTimeClosed | ||||
| 		s.MaxOpenConnections += st.MaxOpenConnections | ||||
| 		s.OpenConnections += st.OpenConnections | ||||
| 		s.WaitCount += st.WaitCount | ||||
| 		s.WaitDuration += st.WaitDuration | ||||
| 		return false | ||||
| 	}) | ||||
| 	return s | ||||
| } | ||||
|  | ||||
| func (c *Cluster) getNodeStateCriterion(ctx context.Context) hasql.NodeStateCriterion { | ||||
| 	if v, ok := ctx.Value(nodeStateCriterionKey{}).(hasql.NodeStateCriterion); ok { | ||||
| 		return v | ||||
| 	} | ||||
| 	return c.options.NodeStateCriterion | ||||
| } | ||||
							
								
								
									
										171
									
								
								cluster/sql/cluster_test.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										171
									
								
								cluster/sql/cluster_test.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,171 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"fmt" | ||||
| 	"testing" | ||||
| 	"time" | ||||
|  | ||||
| 	"github.com/DATA-DOG/go-sqlmock" | ||||
| 	"golang.yandex/hasql/v2" | ||||
| ) | ||||
|  | ||||
| func TestNewCluster(t *testing.T) { | ||||
| 	dbMaster, dbMasterMock, err := sqlmock.New(sqlmock.MonitorPingsOption(true)) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer dbMaster.Close() | ||||
| 	dbMasterMock.MatchExpectationsInOrder(false) | ||||
|  | ||||
| 	dbMasterMock.ExpectQuery(`.*pg_is_in_recovery.*`).WillReturnRows( | ||||
| 		sqlmock.NewRowsWithColumnDefinition( | ||||
| 			sqlmock.NewColumn("role").OfType("int8", 0), | ||||
| 			sqlmock.NewColumn("replication_lag").OfType("int8", 0)). | ||||
| 			AddRow(1, 0)). | ||||
| 		RowsWillBeClosed(). | ||||
| 		WithoutArgs() | ||||
|  | ||||
| 	dbMasterMock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("master-dc1")) | ||||
|  | ||||
| 	dbDRMaster, dbDRMasterMock, err := sqlmock.New(sqlmock.MonitorPingsOption(true)) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer dbDRMaster.Close() | ||||
| 	dbDRMasterMock.MatchExpectationsInOrder(false) | ||||
|  | ||||
| 	dbDRMasterMock.ExpectQuery(`.*pg_is_in_recovery.*`).WillReturnRows( | ||||
| 		sqlmock.NewRowsWithColumnDefinition( | ||||
| 			sqlmock.NewColumn("role").OfType("int8", 0), | ||||
| 			sqlmock.NewColumn("replication_lag").OfType("int8", 0)). | ||||
| 			AddRow(2, 40)). | ||||
| 		RowsWillBeClosed(). | ||||
| 		WithoutArgs() | ||||
|  | ||||
| 	dbDRMasterMock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("drmaster1-dc2")) | ||||
|  | ||||
| 	dbDRMasterMock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("drmaster")) | ||||
|  | ||||
| 	dbSlaveDC1, dbSlaveDC1Mock, err := sqlmock.New(sqlmock.MonitorPingsOption(true)) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer dbSlaveDC1.Close() | ||||
| 	dbSlaveDC1Mock.MatchExpectationsInOrder(false) | ||||
|  | ||||
| 	dbSlaveDC1Mock.ExpectQuery(`.*pg_is_in_recovery.*`).WillReturnRows( | ||||
| 		sqlmock.NewRowsWithColumnDefinition( | ||||
| 			sqlmock.NewColumn("role").OfType("int8", 0), | ||||
| 			sqlmock.NewColumn("replication_lag").OfType("int8", 0)). | ||||
| 			AddRow(2, 50)). | ||||
| 		RowsWillBeClosed(). | ||||
| 		WithoutArgs() | ||||
|  | ||||
| 	dbSlaveDC1Mock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("slave-dc1")) | ||||
|  | ||||
| 	dbSlaveDC2, dbSlaveDC2Mock, err := sqlmock.New(sqlmock.MonitorPingsOption(true)) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer dbSlaveDC2.Close() | ||||
| 	dbSlaveDC1Mock.MatchExpectationsInOrder(false) | ||||
|  | ||||
| 	dbSlaveDC2Mock.ExpectQuery(`.*pg_is_in_recovery.*`).WillReturnRows( | ||||
| 		sqlmock.NewRowsWithColumnDefinition( | ||||
| 			sqlmock.NewColumn("role").OfType("int8", 0), | ||||
| 			sqlmock.NewColumn("replication_lag").OfType("int8", 0)). | ||||
| 			AddRow(2, 50)). | ||||
| 		RowsWillBeClosed(). | ||||
| 		WithoutArgs() | ||||
|  | ||||
| 	dbSlaveDC2Mock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("slave-dc1")) | ||||
|  | ||||
| 	tctx, cancel := context.WithTimeout(t.Context(), 10*time.Second) | ||||
| 	defer cancel() | ||||
|  | ||||
| 	c, err := NewCluster[Querier]( | ||||
| 		WithClusterContext(tctx), | ||||
| 		WithClusterNodeChecker(hasql.PostgreSQLChecker), | ||||
| 		WithClusterNodePicker(NewCustomPicker[Querier]( | ||||
| 			CustomPickerMaxLag(100), | ||||
| 		)), | ||||
| 		WithClusterNodes( | ||||
| 			ClusterNode{"slave-dc1", dbSlaveDC1, 1}, | ||||
| 			ClusterNode{"master-dc1", dbMaster, 1}, | ||||
| 			ClusterNode{"slave-dc2", dbSlaveDC2, 2}, | ||||
| 			ClusterNode{"drmaster1-dc2", dbDRMaster, 0}, | ||||
| 		), | ||||
| 		WithClusterOptions( | ||||
| 			hasql.WithUpdateInterval[Querier](2*time.Second), | ||||
| 			hasql.WithUpdateTimeout[Querier](1*time.Second), | ||||
| 		), | ||||
| 	) | ||||
| 	if err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	defer c.Close() | ||||
|  | ||||
| 	if err = c.WaitForNodes(tctx, hasql.Primary, hasql.Standby); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	time.Sleep(500 * time.Millisecond) | ||||
|  | ||||
| 	node1Name := "" | ||||
| 	fmt.Printf("check for Standby\n") | ||||
| 	if row := c.QueryRowContext(NodeStateCriterion(tctx, hasql.Standby), "SELECT node_name as name"); row.Err() != nil { | ||||
| 		t.Fatal(row.Err()) | ||||
| 	} else if err = row.Scan(&node1Name); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} else if "slave-dc1" != node1Name { | ||||
| 		t.Fatalf("invalid node name %s != %s", "slave-dc1", node1Name) | ||||
| 	} | ||||
|  | ||||
| 	dbSlaveDC1Mock.ExpectQuery(`SELECT node_name as name`).WillReturnRows( | ||||
| 		sqlmock.NewRows([]string{"name"}). | ||||
| 			AddRow("slave-dc1")) | ||||
|  | ||||
| 	node2Name := "" | ||||
| 	fmt.Printf("check for PreferStandby\n") | ||||
| 	if row := c.QueryRowContext(NodeStateCriterion(tctx, hasql.PreferStandby), "SELECT node_name as name"); row.Err() != nil { | ||||
| 		t.Fatal(row.Err()) | ||||
| 	} else if err = row.Scan(&node2Name); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} else if "slave-dc1" != node2Name { | ||||
| 		t.Fatalf("invalid node name %s != %s", "slave-dc1", node2Name) | ||||
| 	} | ||||
|  | ||||
| 	node3Name := "" | ||||
| 	fmt.Printf("check for PreferPrimary\n") | ||||
| 	if row := c.QueryRowContext(NodeStateCriterion(tctx, hasql.PreferPrimary), "SELECT node_name as name"); row.Err() != nil { | ||||
| 		t.Fatal(row.Err()) | ||||
| 	} else if err = row.Scan(&node3Name); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} else if "master-dc1" != node3Name { | ||||
| 		t.Fatalf("invalid node name %s != %s", "master-dc1", node3Name) | ||||
| 	} | ||||
|  | ||||
| 	dbSlaveDC1Mock.ExpectQuery(`.*`).WillReturnRows(sqlmock.NewRows([]string{"role"}).RowError(1, fmt.Errorf("row error"))) | ||||
|  | ||||
| 	time.Sleep(2 * time.Second) | ||||
|  | ||||
| 	fmt.Printf("check for PreferStandby\n") | ||||
| 	if row := c.QueryRowContext(NodeStateCriterion(tctx, hasql.PreferStandby), "SELECT node_name as name"); row.Err() == nil { | ||||
| 		t.Fatal("must return error") | ||||
| 	} | ||||
|  | ||||
| 	if dbMasterErr := dbMasterMock.ExpectationsWereMet(); dbMasterErr != nil { | ||||
| 		t.Error(dbMasterErr) | ||||
| 	} | ||||
| } | ||||
| @@ -1,19 +1,10 @@ | ||||
| // Package codec is an interface for encoding messages | ||||
| package codec // import "go.unistack.org/micro/v3/codec" | ||||
| package codec | ||||
|  | ||||
| import ( | ||||
| 	"errors" | ||||
| 	"io" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/metadata" | ||||
| ) | ||||
|  | ||||
| // Message types | ||||
| const ( | ||||
| 	Error MessageType = iota | ||||
| 	Request | ||||
| 	Response | ||||
| 	Event | ||||
| 	"gopkg.in/yaml.v3" | ||||
| ) | ||||
|  | ||||
| var ( | ||||
| @@ -24,65 +15,23 @@ var ( | ||||
| ) | ||||
|  | ||||
| var ( | ||||
| 	// DefaultMaxMsgSize specifies how much data codec can handle | ||||
| 	DefaultMaxMsgSize = 1024 * 1024 * 4 // 4Mb | ||||
| 	// DefaultCodec is the global default codec | ||||
| 	DefaultCodec = NewCodec() | ||||
| 	// DefaultTagName specifies struct tag name to control codec Marshal/Unmarshal | ||||
| 	DefaultTagName = "codec" | ||||
| ) | ||||
|  | ||||
| // MessageType specifies message type for codec | ||||
| type MessageType int | ||||
|  | ||||
| // Codec encodes/decodes various types of messages used within micro. | ||||
| // ReadHeader and ReadBody are called in pairs to read requests/responses | ||||
| // from the connection. Close is called when finished with the | ||||
| // connection. ReadBody may be called with a nil argument to force the | ||||
| // body to be read and discarded. | ||||
| // Codec encodes/decodes various types of messages. | ||||
| type Codec interface { | ||||
| 	ReadHeader(r io.Reader, m *Message, mt MessageType) error | ||||
| 	ReadBody(r io.Reader, v interface{}) error | ||||
| 	Write(w io.Writer, m *Message, v interface{}) error | ||||
| 	Marshal(v interface{}, opts ...Option) ([]byte, error) | ||||
| 	Unmarshal(b []byte, v interface{}, opts ...Option) error | ||||
| 	String() string | ||||
| } | ||||
|  | ||||
| // Message represents detailed information about | ||||
| // the communication, likely followed by the body. | ||||
| // In the case of an error, body may be nil. | ||||
| type Message struct { | ||||
| 	Header   metadata.Metadata | ||||
| 	Target   string | ||||
| 	Method   string | ||||
| 	Endpoint string | ||||
| 	Error    string | ||||
| 	ID       string | ||||
| 	Body     []byte | ||||
| 	Type     MessageType | ||||
| } | ||||
|  | ||||
| // NewMessage creates new codec message | ||||
| func NewMessage(t MessageType) *Message { | ||||
| 	return &Message{Type: t, Header: metadata.New(0)} | ||||
| } | ||||
|  | ||||
| // MarshalAppend calls codec.Marshal(v) and returns the data appended to buf. | ||||
| // If codec implements MarshalAppend, that is called instead. | ||||
| func MarshalAppend(buf []byte, c Codec, v interface{}, opts ...Option) ([]byte, error) { | ||||
| 	if nc, ok := c.(interface { | ||||
| 		MarshalAppend([]byte, interface{}, ...Option) ([]byte, error) | ||||
| 	}); ok { | ||||
| 		return nc.MarshalAppend(buf, v, opts...) | ||||
| 	} | ||||
|  | ||||
| 	mbuf, err := c.Marshal(v, opts...) | ||||
| 	if err != nil { | ||||
| 		return nil, err | ||||
| 	} | ||||
|  | ||||
| 	return append(buf, mbuf...), nil | ||||
| type CodecV2 interface { | ||||
| 	Marshal(buf []byte, v interface{}, opts ...Option) ([]byte, error) | ||||
| 	Unmarshal(buf []byte, v interface{}, opts ...Option) error | ||||
| 	String() string | ||||
| } | ||||
|  | ||||
| // RawMessage is a raw encoded JSON value. | ||||
| @@ -93,6 +42,8 @@ type RawMessage []byte | ||||
| func (m *RawMessage) MarshalJSON() ([]byte, error) { | ||||
| 	if m == nil { | ||||
| 		return []byte("null"), nil | ||||
| 	} else if len(*m) == 0 { | ||||
| 		return []byte("null"), nil | ||||
| 	} | ||||
| 	return *m, nil | ||||
| } | ||||
| @@ -105,3 +56,22 @@ func (m *RawMessage) UnmarshalJSON(data []byte) error { | ||||
| 	*m = append((*m)[0:0], data...) | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| // MarshalYAML returns m as the JSON encoding of m. | ||||
| func (m *RawMessage) MarshalYAML() ([]byte, error) { | ||||
| 	if m == nil { | ||||
| 		return []byte("null"), nil | ||||
| 	} else if len(*m) == 0 { | ||||
| 		return []byte("null"), nil | ||||
| 	} | ||||
| 	return *m, nil | ||||
| } | ||||
|  | ||||
| // UnmarshalYAML sets *m to a copy of data. | ||||
| func (m *RawMessage) UnmarshalYAML(n *yaml.Node) error { | ||||
| 	if m == nil { | ||||
| 		return errors.New("RawMessage UnmarshalYAML on nil pointer") | ||||
| 	} | ||||
| 	*m = append((*m)[0:0], []byte(n.Value)...) | ||||
| 	return nil | ||||
| } | ||||
|   | ||||
| @@ -15,6 +15,15 @@ func FromContext(ctx context.Context) (Codec, bool) { | ||||
| 	return c, ok | ||||
| } | ||||
|  | ||||
| // MustContext returns codec from context | ||||
| func MustContext(ctx context.Context) Codec { | ||||
| 	c, ok := FromContext(ctx) | ||||
| 	if !ok { | ||||
| 		panic("missing codec") | ||||
| 	} | ||||
| 	return c | ||||
| } | ||||
|  | ||||
| // NewContext put codec in context | ||||
| func NewContext(ctx context.Context, c Codec) context.Context { | ||||
| 	if ctx == nil { | ||||
|   | ||||
| @@ -1,5 +1,7 @@ | ||||
| package codec | ||||
|  | ||||
| import "gopkg.in/yaml.v3" | ||||
|  | ||||
| // Frame gives us the ability to define raw data to send over the pipes | ||||
| type Frame struct { | ||||
| 	Data []byte | ||||
| @@ -20,6 +22,17 @@ func (m *Frame) UnmarshalJSON(data []byte) error { | ||||
| 	return m.Unmarshal(data) | ||||
| } | ||||
|  | ||||
| // MarshalYAML returns frame data | ||||
| func (m *Frame) MarshalYAML() ([]byte, error) { | ||||
| 	return m.Marshal() | ||||
| } | ||||
|  | ||||
| // UnmarshalYAML set frame data | ||||
| func (m *Frame) UnmarshalYAML(n *yaml.Node) error { | ||||
| 	m.Data = []byte(n.Value) | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| // ProtoMessage noop func | ||||
| func (m *Frame) ProtoMessage() {} | ||||
|  | ||||
|   | ||||
| @@ -2,70 +2,14 @@ package codec | ||||
|  | ||||
| import ( | ||||
| 	"encoding/json" | ||||
| 	"io" | ||||
|  | ||||
| 	codecpb "go.unistack.org/micro-proto/v3/codec" | ||||
| ) | ||||
|  | ||||
| type noopCodec struct { | ||||
| 	opts Options | ||||
| } | ||||
|  | ||||
| func (c *noopCodec) ReadHeader(conn io.Reader, m *Message, t MessageType) error { | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| func (c *noopCodec) ReadBody(conn io.Reader, b interface{}) error { | ||||
| 	// read bytes | ||||
| 	buf, err := io.ReadAll(conn) | ||||
| 	if err != nil { | ||||
| 		return err | ||||
| 	} | ||||
|  | ||||
| 	if b == nil { | ||||
| 		return nil | ||||
| 	} | ||||
|  | ||||
| 	switch v := b.(type) { | ||||
| 	case *string: | ||||
| 		*v = string(buf) | ||||
| 	case *[]byte: | ||||
| 		*v = buf | ||||
| 	case *Frame: | ||||
| 		v.Data = buf | ||||
| 	default: | ||||
| 		return json.Unmarshal(buf, v) | ||||
| 	} | ||||
|  | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| func (c *noopCodec) Write(conn io.Writer, m *Message, b interface{}) error { | ||||
| 	if b == nil { | ||||
| 		return nil | ||||
| 	} | ||||
|  | ||||
| 	var v []byte | ||||
| 	switch vb := b.(type) { | ||||
| 	case *Frame: | ||||
| 		v = vb.Data | ||||
| 	case string: | ||||
| 		v = []byte(vb) | ||||
| 	case *string: | ||||
| 		v = []byte(*vb) | ||||
| 	case *[]byte: | ||||
| 		v = *vb | ||||
| 	case []byte: | ||||
| 		v = vb | ||||
| 	default: | ||||
| 		var err error | ||||
| 		v, err = json.Marshal(vb) | ||||
| 		if err != nil { | ||||
| 			return err | ||||
| 		} | ||||
| 	} | ||||
| 	_, err := conn.Write(v) | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| func (c *noopCodec) String() string { | ||||
| 	return "noop" | ||||
| } | ||||
| @@ -91,8 +35,8 @@ func (c *noopCodec) Marshal(v interface{}, opts ...Option) ([]byte, error) { | ||||
| 		return ve, nil | ||||
| 	case *Frame: | ||||
| 		return ve.Data, nil | ||||
| 	case *Message: | ||||
| 		return ve.Body, nil | ||||
| 	case *codecpb.Frame: | ||||
| 		return ve.Data, nil | ||||
| 	} | ||||
|  | ||||
| 	return json.Marshal(v) | ||||
| @@ -115,8 +59,8 @@ func (c *noopCodec) Unmarshal(d []byte, v interface{}, opts ...Option) error { | ||||
| 	case *Frame: | ||||
| 		ve.Data = d | ||||
| 		return nil | ||||
| 	case *Message: | ||||
| 		ve.Body = d | ||||
| 	case *codecpb.Frame: | ||||
| 		ve.Data = d | ||||
| 		return nil | ||||
| 	} | ||||
|  | ||||
|   | ||||
| @@ -23,15 +23,8 @@ type Options struct { | ||||
| 	Context context.Context | ||||
| 	// TagName specifies tag name in struct to control codec | ||||
| 	TagName string | ||||
| 	// MaxMsgSize specifies max messages size that reads by codec | ||||
| 	MaxMsgSize int | ||||
| } | ||||
|  | ||||
| // MaxMsgSize sets the max message size | ||||
| func MaxMsgSize(n int) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.MaxMsgSize = n | ||||
| 	} | ||||
| 	// Flatten specifies that struct must be analyzed for flatten tag | ||||
| 	Flatten bool | ||||
| } | ||||
|  | ||||
| // TagName sets the codec tag name in struct | ||||
| @@ -41,6 +34,13 @@ func TagName(n string) Option { | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Flatten enables checking for flatten tag name | ||||
| func Flatten(b bool) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Flatten = b | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Logger sets the logger | ||||
| func Logger(l logger.Logger) Option { | ||||
| 	return func(o *Options) { | ||||
| @@ -65,12 +65,12 @@ func Meter(m meter.Meter) Option { | ||||
| // NewOptions returns new options | ||||
| func NewOptions(opts ...Option) Options { | ||||
| 	options := Options{ | ||||
| 		Context:    context.Background(), | ||||
| 		Logger:     logger.DefaultLogger, | ||||
| 		Meter:      meter.DefaultMeter, | ||||
| 		Tracer:     tracer.DefaultTracer, | ||||
| 		MaxMsgSize: DefaultMaxMsgSize, | ||||
| 		TagName:    DefaultTagName, | ||||
| 		Context: context.Background(), | ||||
| 		Logger:  logger.DefaultLogger, | ||||
| 		Meter:   meter.DefaultMeter, | ||||
| 		Tracer:  tracer.DefaultTracer, | ||||
| 		TagName: DefaultTagName, | ||||
| 		Flatten: false, | ||||
| 	} | ||||
|  | ||||
| 	for _, o := range opts { | ||||
|   | ||||
| @@ -1,5 +1,5 @@ | ||||
| // Package config is an interface for dynamic configuration. | ||||
| package config // import "go.unistack.org/micro/v3/config" | ||||
| package config | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| @@ -13,7 +13,7 @@ type Validator interface { | ||||
| } | ||||
|  | ||||
| // DefaultConfig default config | ||||
| var DefaultConfig Config = NewConfig() | ||||
| var DefaultConfig = NewConfig() | ||||
|  | ||||
| // DefaultWatcherMinInterval default min interval for poll changes | ||||
| var DefaultWatcherMinInterval = 5 * time.Second | ||||
| @@ -50,6 +50,13 @@ type Config interface { | ||||
| 	String() string | ||||
| } | ||||
|  | ||||
| type ( | ||||
| 	FuncLoad func(ctx context.Context, opts ...LoadOption) error | ||||
| 	HookLoad func(next FuncLoad) FuncLoad | ||||
| 	FuncSave func(ctx context.Context, opts ...SaveOption) error | ||||
| 	HookSave func(next FuncSave) FuncSave | ||||
| ) | ||||
|  | ||||
| // Watcher is the config watcher | ||||
| type Watcher interface { | ||||
| 	// Next blocks until update happens or error returned | ||||
| @@ -131,7 +138,7 @@ var ( | ||||
| 				return nil | ||||
| 			} | ||||
| 			if err := fn(ctx, c); err != nil { | ||||
| 				c.Options().Logger.Errorf(ctx, "%s BeforeLoad err: %v", c.String(), err) | ||||
| 				c.Options().Logger.Error(ctx, c.String()+" BeforeLoad error", err) | ||||
| 				if !c.Options().AllowFail { | ||||
| 					return err | ||||
| 				} | ||||
| @@ -146,7 +153,7 @@ var ( | ||||
| 				return nil | ||||
| 			} | ||||
| 			if err := fn(ctx, c); err != nil { | ||||
| 				c.Options().Logger.Errorf(ctx, "%s AfterLoad err: %v", c.String(), err) | ||||
| 				c.Options().Logger.Error(ctx, c.String()+" AfterLoad error", err) | ||||
| 				if !c.Options().AllowFail { | ||||
| 					return err | ||||
| 				} | ||||
| @@ -161,7 +168,7 @@ var ( | ||||
| 				return nil | ||||
| 			} | ||||
| 			if err := fn(ctx, c); err != nil { | ||||
| 				c.Options().Logger.Errorf(ctx, "%s BeforeSave err: %v", c.String(), err) | ||||
| 				c.Options().Logger.Error(ctx, c.String()+" BeforeSave error", err) | ||||
| 				if !c.Options().AllowFail { | ||||
| 					return err | ||||
| 				} | ||||
| @@ -176,7 +183,7 @@ var ( | ||||
| 				return nil | ||||
| 			} | ||||
| 			if err := fn(ctx, c); err != nil { | ||||
| 				c.Options().Logger.Errorf(ctx, "%s AfterSave err: %v", c.String(), err) | ||||
| 				c.Options().Logger.Error(ctx, c.String()+" AfterSave error", err) | ||||
| 				if !c.Options().AllowFail { | ||||
| 					return err | ||||
| 				} | ||||
| @@ -191,7 +198,7 @@ var ( | ||||
| 				return nil | ||||
| 			} | ||||
| 			if err := fn(ctx, c); err != nil { | ||||
| 				c.Options().Logger.Errorf(ctx, "%s BeforeInit err: %v", c.String(), err) | ||||
| 				c.Options().Logger.Error(ctx, c.String()+" BeforeInit error", err) | ||||
| 				if !c.Options().AllowFail { | ||||
| 					return err | ||||
| 				} | ||||
| @@ -206,7 +213,7 @@ var ( | ||||
| 				return nil | ||||
| 			} | ||||
| 			if err := fn(ctx, c); err != nil { | ||||
| 				c.Options().Logger.Errorf(ctx, "%s AfterInit err: %v", c.String(), err) | ||||
| 				c.Options().Logger.Error(ctx, c.String()+" AfterInit error", err) | ||||
| 				if !c.Options().AllowFail { | ||||
| 					return err | ||||
| 				} | ||||
|   | ||||
| @@ -15,6 +15,15 @@ func FromContext(ctx context.Context) (Config, bool) { | ||||
| 	return c, ok | ||||
| } | ||||
|  | ||||
| // MustContext returns store from context | ||||
| func MustContext(ctx context.Context) Config { | ||||
| 	c, ok := FromContext(ctx) | ||||
| 	if !ok { | ||||
| 		panic("missing config") | ||||
| 	} | ||||
| 	return c | ||||
| } | ||||
|  | ||||
| // NewContext put store in context | ||||
| func NewContext(ctx context.Context, c Config) context.Context { | ||||
| 	if ctx == nil { | ||||
|   | ||||
| @@ -9,13 +9,16 @@ import ( | ||||
|  | ||||
| 	"dario.cat/mergo" | ||||
| 	"github.com/google/uuid" | ||||
| 	"go.unistack.org/micro/v3/options" | ||||
| 	mid "go.unistack.org/micro/v3/util/id" | ||||
| 	rutil "go.unistack.org/micro/v3/util/reflect" | ||||
| 	mtime "go.unistack.org/micro/v3/util/time" | ||||
| ) | ||||
|  | ||||
| type defaultConfig struct { | ||||
| 	opts Options | ||||
| 	funcLoad FuncLoad | ||||
| 	funcSave FuncSave | ||||
| 	opts     Options | ||||
| } | ||||
|  | ||||
| func (c *defaultConfig) Options() Options { | ||||
| @@ -31,6 +34,18 @@ func (c *defaultConfig) Init(opts ...Option) error { | ||||
| 		return err | ||||
| 	} | ||||
|  | ||||
| 	c.funcLoad = c.fnLoad | ||||
| 	c.funcSave = c.fnSave | ||||
|  | ||||
| 	c.opts.Hooks.EachPrev(func(hook options.Hook) { | ||||
| 		switch h := hook.(type) { | ||||
| 		case HookLoad: | ||||
| 			c.funcLoad = h(c.funcLoad) | ||||
| 		case HookSave: | ||||
| 			c.funcSave = h(c.funcSave) | ||||
| 		} | ||||
| 	}) | ||||
|  | ||||
| 	if err := DefaultAfterInit(c.opts.Context, c); err != nil && !c.opts.AllowFail { | ||||
| 		return err | ||||
| 	} | ||||
| @@ -39,11 +54,17 @@ func (c *defaultConfig) Init(opts ...Option) error { | ||||
| } | ||||
|  | ||||
| func (c *defaultConfig) Load(ctx context.Context, opts ...LoadOption) error { | ||||
| 	return c.funcLoad(ctx, opts...) | ||||
| } | ||||
|  | ||||
| func (c *defaultConfig) fnLoad(ctx context.Context, opts ...LoadOption) error { | ||||
| 	var err error | ||||
|  | ||||
| 	if c.opts.SkipLoad != nil && c.opts.SkipLoad(ctx, c) { | ||||
| 		return nil | ||||
| 	} | ||||
|  | ||||
| 	if err := DefaultBeforeLoad(ctx, c); err != nil && !c.opts.AllowFail { | ||||
| 	if err = DefaultBeforeLoad(ctx, c); err != nil && !c.opts.AllowFail { | ||||
| 		return err | ||||
| 	} | ||||
|  | ||||
| @@ -233,6 +254,7 @@ func fillValue(value reflect.Value, val string) error { | ||||
| 		} | ||||
| 		value.Set(reflect.ValueOf(v)) | ||||
| 	} | ||||
|  | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| @@ -295,7 +317,11 @@ func fillValues(valueOf reflect.Value, tname string) error { | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| func (c *defaultConfig) Save(ctx context.Context, _ ...SaveOption) error { | ||||
| func (c *defaultConfig) Save(ctx context.Context, opts ...SaveOption) error { | ||||
| 	return c.funcSave(ctx, opts...) | ||||
| } | ||||
|  | ||||
| func (c *defaultConfig) fnSave(ctx context.Context, opts ...SaveOption) error { | ||||
| 	if c.opts.SkipSave != nil && c.opts.SkipSave(ctx, c) { | ||||
| 		return nil | ||||
| 	} | ||||
| @@ -319,7 +345,7 @@ func (c *defaultConfig) Name() string { | ||||
| 	return c.opts.Name | ||||
| } | ||||
|  | ||||
| func (c *defaultConfig) Watch(ctx context.Context, opts ...WatchOption) (Watcher, error) { | ||||
| func (c *defaultConfig) Watch(_ context.Context, _ ...WatchOption) (Watcher, error) { | ||||
| 	return nil, ErrWatcherNotImplemented | ||||
| } | ||||
|  | ||||
| @@ -329,5 +355,9 @@ func NewConfig(opts ...Option) Config { | ||||
| 	if len(options.StructTag) == 0 { | ||||
| 		options.StructTag = "default" | ||||
| 	} | ||||
| 	return &defaultConfig{opts: options} | ||||
| 	c := &defaultConfig{opts: options} | ||||
| 	c.funcLoad = c.fnLoad | ||||
| 	c.funcSave = c.fnSave | ||||
|  | ||||
| 	return c | ||||
| } | ||||
|   | ||||
| @@ -3,24 +3,26 @@ package config_test | ||||
| import ( | ||||
| 	"context" | ||||
| 	"fmt" | ||||
| 	"reflect" | ||||
| 	"testing" | ||||
| 	"time" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/config" | ||||
| 	mid "go.unistack.org/micro/v3/util/id" | ||||
| 	mtime "go.unistack.org/micro/v3/util/time" | ||||
| ) | ||||
|  | ||||
| type cfg struct { | ||||
| 	StringValue    string `default:"string_value"` | ||||
| 	IgnoreValue    string `json:"-"` | ||||
| 	StructValue    *cfgStructValue | ||||
| 	IntValue       int             `default:"99"` | ||||
| 	DurationValue  time.Duration   `default:"10s"` | ||||
| 	MDurationValue mtime.Duration  `default:"10s"` | ||||
| 	MapValue       map[string]bool `default:"key1=true,key2=false"` | ||||
| 	UUIDValue      string          `default:"micro:generate uuid"` | ||||
| 	IDValue        string          `default:"micro:generate id"` | ||||
| 	MapValue    map[string]bool `default:"key1=true,key2=false"` | ||||
| 	StructValue *cfgStructValue | ||||
|  | ||||
| 	StringValue string `default:"string_value"` | ||||
| 	IgnoreValue string `json:"-"` | ||||
| 	UUIDValue   string `default:"micro:generate uuid"` | ||||
| 	IDValue     string `default:"micro:generate id"` | ||||
|  | ||||
| 	DurationValue  time.Duration  `default:"10s"` | ||||
| 	MDurationValue mtime.Duration `default:"10s"` | ||||
| 	IntValue       int            `default:"99"` | ||||
| } | ||||
|  | ||||
| type cfgStructValue struct { | ||||
| @@ -41,6 +43,35 @@ func (c *cfgStructValue) Validate() error { | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| type testHook struct { | ||||
| 	f bool | ||||
| } | ||||
|  | ||||
| func (t *testHook) Load(fn config.FuncLoad) config.FuncLoad { | ||||
| 	return func(ctx context.Context, opts ...config.LoadOption) error { | ||||
| 		t.f = true | ||||
| 		return fn(ctx, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestHook(t *testing.T) { | ||||
| 	h := &testHook{} | ||||
|  | ||||
| 	c := config.NewConfig(config.Struct(h), config.Hooks(config.HookLoad(h.Load))) | ||||
|  | ||||
| 	if err := c.Init(); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	if err := c.Load(context.TODO()); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	if !h.f { | ||||
| 		t.Fatal("hook not works") | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestDefault(t *testing.T) { | ||||
| 	ctx := context.Background() | ||||
| 	conf := &cfg{IntValue: 10} | ||||
| @@ -83,8 +114,6 @@ func TestDefault(t *testing.T) { | ||||
|  | ||||
| 	if conf.IDValue == "" { | ||||
| 		t.Fatalf("id value empty") | ||||
| 	} else if len(conf.IDValue) != mid.DefaultSize { | ||||
| 		t.Fatalf("id value invalid: %s", conf.IDValue) | ||||
| 	} | ||||
| 	_ = conf | ||||
| 	// t.Logf("%#+v\n", conf) | ||||
| @@ -105,3 +134,13 @@ func TestValidate(t *testing.T) { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func Test_SizeOf(t *testing.T) { | ||||
| 	st := cfg{} | ||||
|  | ||||
| 	tVal := reflect.TypeOf(st) | ||||
| 	for i := 0; i < tVal.NumField(); i++ { | ||||
| 		field := tVal.Field(i) | ||||
| 		fmt.Printf("Field: %s, Offset: %d, Size: %d\n", field.Name, field.Offset, field.Type.Size()) | ||||
| 	} | ||||
| } | ||||
|   | ||||
| @@ -7,6 +7,7 @@ import ( | ||||
| 	"go.unistack.org/micro/v3/codec" | ||||
| 	"go.unistack.org/micro/v3/logger" | ||||
| 	"go.unistack.org/micro/v3/meter" | ||||
| 	"go.unistack.org/micro/v3/options" | ||||
| 	"go.unistack.org/micro/v3/tracer" | ||||
| ) | ||||
|  | ||||
| @@ -40,12 +41,16 @@ type Options struct { | ||||
| 	BeforeInit []func(context.Context, Config) error | ||||
| 	// AfterInit contains slice of funcs that runs after Init | ||||
| 	AfterInit []func(context.Context, Config) error | ||||
| 	// AllowFail flag to allow fail in config source | ||||
| 	AllowFail bool | ||||
|  | ||||
| 	// SkipLoad runs only if condition returns true | ||||
| 	SkipLoad func(context.Context, Config) bool | ||||
| 	// SkipSave runs only if condition returns true | ||||
| 	SkipSave func(context.Context, Config) bool | ||||
| 	// Hooks can be run before/after config Save/Load | ||||
| 	Hooks options.Hooks | ||||
|  | ||||
| 	// AllowFail flag to allow fail in config source | ||||
| 	AllowFail bool | ||||
| } | ||||
|  | ||||
| // Option function signature | ||||
| @@ -275,10 +280,10 @@ func WatchCoalesce(b bool) WatchOption { | ||||
| } | ||||
|  | ||||
| // WatchInterval specifies min and max time.Duration for pulling changes | ||||
| func WatchInterval(min, max time.Duration) WatchOption { | ||||
| func WatchInterval(minTime, maxTime time.Duration) WatchOption { | ||||
| 	return func(o *WatchOptions) { | ||||
| 		o.MinInterval = min | ||||
| 		o.MaxInterval = max | ||||
| 		o.MinInterval = minTime | ||||
| 		o.MaxInterval = maxTime | ||||
| 	} | ||||
| } | ||||
|  | ||||
| @@ -288,3 +293,10 @@ func WatchStruct(src interface{}) WatchOption { | ||||
| 		o.Struct = src | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Hooks sets hook runs before action | ||||
| func Hooks(h ...options.Hook) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Hooks = append(o.Hooks, h...) | ||||
| 	} | ||||
| } | ||||
|   | ||||
| @@ -1,6 +1,6 @@ | ||||
| // Package errors provides a way to return detailed information | ||||
| // for an RPC request error. The error is normally JSON encoded. | ||||
| package errors // import "go.unistack.org/micro/v3/errors" | ||||
| package errors | ||||
|  | ||||
| import ( | ||||
| 	"bytes" | ||||
| @@ -44,6 +44,20 @@ var ( | ||||
| 	ErrGatewayTimeout = &Error{Code: 504} | ||||
| ) | ||||
|  | ||||
| const ProblemContentType = "application/problem+json" | ||||
|  | ||||
| type Problem struct { | ||||
| 	Type     string `json:"type,omitempty"` | ||||
| 	Title    string `json:"title,omitempty"` | ||||
| 	Detail   string `json:"detail,omitempty"` | ||||
| 	Instance string `json:"instance,omitempty"` | ||||
| 	Errors   []struct { | ||||
| 		Title  string `json:"title,omitempty"` | ||||
| 		Detail string `json:"detail,omitempty"` | ||||
| 	} `json:"errors,omitempty"` | ||||
| 	Status int `json:"status,omitempty"` | ||||
| } | ||||
|  | ||||
| // Error type | ||||
| type Error struct { | ||||
| 	// ID holds error id or service, usually someting like my_service or id | ||||
| @@ -262,6 +276,10 @@ func CodeIn(err interface{}, codes ...int32) bool { | ||||
|  | ||||
| // FromError try to convert go error to *Error | ||||
| func FromError(err error) *Error { | ||||
| 	if err == nil { | ||||
| 		return nil | ||||
| 	} | ||||
|  | ||||
| 	if verr, ok := err.(*Error); ok && verr != nil { | ||||
| 		return verr | ||||
| 	} | ||||
|   | ||||
| @@ -2,7 +2,7 @@ package errors | ||||
|  | ||||
| import ( | ||||
| 	"encoding/json" | ||||
| 	er "errors" | ||||
| 	"errors" | ||||
| 	"fmt" | ||||
| 	"net/http" | ||||
| 	"testing" | ||||
| @@ -26,7 +26,7 @@ func TestMarshalJSON(t *testing.T) { | ||||
| func TestEmpty(t *testing.T) { | ||||
| 	msg := "test" | ||||
| 	var err *Error | ||||
| 	err = FromError(fmt.Errorf(msg)) | ||||
| 	err = FromError(errors.New(msg)) | ||||
| 	if err.Detail != msg { | ||||
| 		t.Fatalf("invalid error %v", err) | ||||
| 	} | ||||
| @@ -42,7 +42,7 @@ func TestFromError(t *testing.T) { | ||||
| 	if merr.ID != "go.micro.test" || merr.Code != 404 { | ||||
| 		t.Fatalf("invalid conversation %v != %v", err, merr) | ||||
| 	} | ||||
| 	err = er.New(err.Error()) | ||||
| 	err = errors.New(err.Error()) | ||||
| 	merr = FromError(err) | ||||
| 	if merr.ID != "go.micro.test" || merr.Code != 404 { | ||||
| 		t.Fatalf("invalid conversation %v != %v", err, merr) | ||||
| @@ -57,7 +57,7 @@ func TestEqual(t *testing.T) { | ||||
| 		t.Fatal("errors must be equal") | ||||
| 	} | ||||
|  | ||||
| 	err3 := er.New("my test err") | ||||
| 	err3 := errors.New("my test err") | ||||
| 	if Equal(err1, err3) { | ||||
| 		t.Fatal("errors must be not equal") | ||||
| 	} | ||||
|   | ||||
| @@ -15,6 +15,15 @@ func FromContext(ctx context.Context) (Flow, bool) { | ||||
| 	return c, ok | ||||
| } | ||||
|  | ||||
| // MustContext returns Flow from context | ||||
| func MustContext(ctx context.Context) Flow { | ||||
| 	f, ok := FromContext(ctx) | ||||
| 	if !ok { | ||||
| 		panic("missing flow") | ||||
| 	} | ||||
| 	return f | ||||
| } | ||||
|  | ||||
| // NewContext stores Flow to context | ||||
| func NewContext(ctx context.Context, f Flow) context.Context { | ||||
| 	if ctx == nil { | ||||
|   | ||||
| @@ -188,7 +188,7 @@ func (w *microWorkflow) Execute(ctx context.Context, req *Message, opts ...Execu | ||||
| 	steps, err := w.getSteps(options.Start, options.Reverse) | ||||
| 	if err != nil { | ||||
| 		if werr := workflowStore.Write(w.opts.Context, "status", &codec.Frame{Data: []byte(StatusPending.String())}); werr != nil { | ||||
| 			w.opts.Logger.Errorf(w.opts.Context, "store error: %v", werr) | ||||
| 			w.opts.Logger.Error(w.opts.Context, "store write error", werr) | ||||
| 		} | ||||
| 		return "", err | ||||
| 	} | ||||
| @@ -212,7 +212,7 @@ func (w *microWorkflow) Execute(ctx context.Context, req *Message, opts ...Execu | ||||
| 	done := make(chan struct{}) | ||||
|  | ||||
| 	if werr := workflowStore.Write(w.opts.Context, "status", &codec.Frame{Data: []byte(StatusRunning.String())}); werr != nil { | ||||
| 		w.opts.Logger.Errorf(w.opts.Context, "store error: %v", werr) | ||||
| 		w.opts.Logger.Error(w.opts.Context, "store write error", werr) | ||||
| 		return eid, werr | ||||
| 	} | ||||
| 	for idx := range steps { | ||||
| @@ -237,7 +237,7 @@ func (w *microWorkflow) Execute(ctx context.Context, req *Message, opts ...Execu | ||||
| 					return | ||||
| 				} | ||||
| 				if w.opts.Logger.V(logger.TraceLevel) { | ||||
| 					w.opts.Logger.Tracef(nctx, "will be executed %v", steps[idx][nidx]) | ||||
| 					w.opts.Logger.Trace(nctx, fmt.Sprintf("will be executed %v", steps[idx][nidx])) | ||||
| 				} | ||||
| 				cstep := steps[idx][nidx] | ||||
| 				// nolint: nestif | ||||
| @@ -257,21 +257,21 @@ func (w *microWorkflow) Execute(ctx context.Context, req *Message, opts ...Execu | ||||
| 						if serr != nil { | ||||
| 							step.SetStatus(StatusFailure) | ||||
| 							if werr := stepStore.Write(ctx, step.ID()+w.opts.Store.Options().Separator+"rsp", serr); werr != nil && w.opts.Logger.V(logger.ErrorLevel) { | ||||
| 								w.opts.Logger.Errorf(ctx, "store write error: %v", werr) | ||||
| 								w.opts.Logger.Error(ctx, "store write error", werr) | ||||
| 							} | ||||
| 							if werr := stepStore.Write(ctx, step.ID()+w.opts.Store.Options().Separator+"status", &codec.Frame{Data: []byte(StatusFailure.String())}); werr != nil && w.opts.Logger.V(logger.ErrorLevel) { | ||||
| 								w.opts.Logger.Errorf(ctx, "store write error: %v", werr) | ||||
| 								w.opts.Logger.Error(ctx, "store write error", werr) | ||||
| 							} | ||||
| 							cherr <- serr | ||||
| 							return | ||||
| 						} | ||||
| 						if werr := stepStore.Write(ctx, step.ID()+w.opts.Store.Options().Separator+"rsp", rsp); werr != nil { | ||||
| 							w.opts.Logger.Errorf(ctx, "store write error: %v", werr) | ||||
| 							w.opts.Logger.Error(ctx, "store write error", werr) | ||||
| 							cherr <- werr | ||||
| 							return | ||||
| 						} | ||||
| 						if werr := stepStore.Write(ctx, step.ID()+w.opts.Store.Options().Separator+"status", &codec.Frame{Data: []byte(StatusSuccess.String())}); werr != nil { | ||||
| 							w.opts.Logger.Errorf(ctx, "store write error: %v", werr) | ||||
| 							w.opts.Logger.Error(ctx, "store write error", werr) | ||||
| 							cherr <- werr | ||||
| 							return | ||||
| 						} | ||||
| @@ -290,16 +290,16 @@ func (w *microWorkflow) Execute(ctx context.Context, req *Message, opts ...Execu | ||||
| 					if serr != nil { | ||||
| 						cstep.SetStatus(StatusFailure) | ||||
| 						if werr := stepStore.Write(ctx, cstep.ID()+w.opts.Store.Options().Separator+"rsp", serr); werr != nil && w.opts.Logger.V(logger.ErrorLevel) { | ||||
| 							w.opts.Logger.Errorf(ctx, "store write error: %v", werr) | ||||
| 							w.opts.Logger.Error(ctx, "store write error", werr) | ||||
| 						} | ||||
| 						if werr := stepStore.Write(ctx, cstep.ID()+w.opts.Store.Options().Separator+"status", &codec.Frame{Data: []byte(StatusFailure.String())}); werr != nil && w.opts.Logger.V(logger.ErrorLevel) { | ||||
| 							w.opts.Logger.Errorf(ctx, "store write error: %v", werr) | ||||
| 							w.opts.Logger.Error(ctx, "store write error", werr) | ||||
| 						} | ||||
| 						cherr <- serr | ||||
| 						return | ||||
| 					} | ||||
| 					if werr := stepStore.Write(ctx, cstep.ID()+w.opts.Store.Options().Separator+"rsp", rsp); werr != nil { | ||||
| 						w.opts.Logger.Errorf(ctx, "store write error: %v", werr) | ||||
| 						w.opts.Logger.Error(ctx, "store write error", werr) | ||||
| 						cherr <- werr | ||||
| 						return | ||||
| 					} | ||||
| @@ -317,7 +317,7 @@ func (w *microWorkflow) Execute(ctx context.Context, req *Message, opts ...Execu | ||||
| 		return eid, nil | ||||
| 	} | ||||
|  | ||||
| 	logger.Tracef(ctx, "wait for finish or error") | ||||
| 	logger.DefaultLogger.Trace(ctx, "wait for finish or error") | ||||
| 	select { | ||||
| 	case <-nctx.Done(): | ||||
| 		err = nctx.Err() | ||||
| @@ -333,15 +333,15 @@ func (w *microWorkflow) Execute(ctx context.Context, req *Message, opts ...Execu | ||||
| 	switch { | ||||
| 	case nctx.Err() != nil: | ||||
| 		if werr := workflowStore.Write(w.opts.Context, "status", &codec.Frame{Data: []byte(StatusAborted.String())}); werr != nil { | ||||
| 			w.opts.Logger.Errorf(w.opts.Context, "store error: %v", werr) | ||||
| 			w.opts.Logger.Error(w.opts.Context, "store write error", werr) | ||||
| 		} | ||||
| 	case err == nil: | ||||
| 		if werr := workflowStore.Write(w.opts.Context, "status", &codec.Frame{Data: []byte(StatusSuccess.String())}); werr != nil { | ||||
| 			w.opts.Logger.Errorf(w.opts.Context, "store error: %v", werr) | ||||
| 			w.opts.Logger.Error(w.opts.Context, "store write error", werr) | ||||
| 		} | ||||
| 	case err != nil: | ||||
| 		if werr := workflowStore.Write(w.opts.Context, "status", &codec.Frame{Data: []byte(StatusFailure.String())}); werr != nil { | ||||
| 			w.opts.Logger.Errorf(w.opts.Context, "store error: %v", werr) | ||||
| 			w.opts.Logger.Error(w.opts.Context, "store write error", werr) | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
|   | ||||
| @@ -1,5 +1,5 @@ | ||||
| // Package flow is an interface used for saga pattern microservice workflow | ||||
| package flow // import "go.unistack.org/micro/v3/flow" | ||||
| package flow | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
|   | ||||
| @@ -32,7 +32,7 @@ type fsm struct { | ||||
|  | ||||
| // NewFSM creates a new finite state machine having the specified initial state | ||||
| // with specified options | ||||
| func NewFSM(opts ...Option) *fsm { | ||||
| func NewFSM(opts ...Option) FSM { | ||||
| 	return &fsm{ | ||||
| 		statesMap: map[string]StateFunc{}, | ||||
| 		opts:      NewOptions(opts...), | ||||
|   | ||||
| @@ -1,4 +1,4 @@ | ||||
| package fsm // import "go.unistack.org/micro/v3/fsm" | ||||
| package fsm | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
|   | ||||
| @@ -17,7 +17,7 @@ func TestFSMStart(t *testing.T) { | ||||
|  | ||||
| 	wrapper := func(next StateFunc) StateFunc { | ||||
| 		return func(sctx context.Context, s State, opts ...StateOption) (State, error) { | ||||
| 			sctx = logger.NewContext(sctx, logger.Fields("state", s.Name())) | ||||
| 			sctx = logger.NewContext(sctx, logger.DefaultLogger.Fields("state", s.Name())) | ||||
| 			return next(sctx, s, opts...) | ||||
| 		} | ||||
| 	} | ||||
|   | ||||
							
								
								
									
										44
									
								
								go.mod
									
									
									
									
									
								
							
							
						
						
									
										44
									
								
								go.mod
									
									
									
									
									
								
							| @@ -1,20 +1,44 @@ | ||||
| module go.unistack.org/micro/v3 | ||||
|  | ||||
| go 1.20 | ||||
| go 1.24.0 | ||||
|  | ||||
| require ( | ||||
| 	dario.cat/mergo v1.0.0 | ||||
| 	github.com/DATA-DOG/go-sqlmock v1.5.0 | ||||
| 	github.com/google/uuid v1.3.0 | ||||
| 	dario.cat/mergo v1.0.1 | ||||
| 	github.com/DATA-DOG/go-sqlmock v1.5.2 | ||||
| 	github.com/KimMachineGun/automemlimit v0.6.1 | ||||
| 	github.com/ash3in/uuidv8 v1.2.0 | ||||
| 	github.com/google/uuid v1.6.0 | ||||
| 	github.com/matoous/go-nanoid v1.5.1 | ||||
| 	github.com/patrickmn/go-cache v2.1.0+incompatible | ||||
| 	github.com/silas/dag v0.0.0-20220518035006-a7e85ada93c5 | ||||
| 	golang.org/x/sync v0.3.0 | ||||
| 	google.golang.org/grpc v1.57.0 | ||||
| 	google.golang.org/protobuf v1.31.0 | ||||
| 	github.com/stretchr/testify v1.10.0 | ||||
| 	go.uber.org/automaxprocs v1.6.0 | ||||
| 	go.unistack.org/micro-proto/v3 v3.4.1 | ||||
| 	golang.org/x/sync v0.10.0 | ||||
| 	golang.yandex/hasql/v2 v2.1.0 | ||||
| 	google.golang.org/grpc v1.69.2 | ||||
| 	google.golang.org/protobuf v1.36.1 | ||||
| 	gopkg.in/yaml.v3 v3.0.1 | ||||
| ) | ||||
|  | ||||
| require ( | ||||
| 	github.com/golang/protobuf v1.5.3 // indirect | ||||
| 	golang.org/x/net v0.14.0 // indirect | ||||
| 	google.golang.org/genproto/googleapis/rpc v0.0.0-20230526203410-71b5a4ffd15e // indirect | ||||
| 	github.com/cilium/ebpf v0.16.0 // indirect | ||||
| 	github.com/containerd/cgroups/v3 v3.0.4 // indirect | ||||
| 	github.com/containerd/log v0.1.0 // indirect | ||||
| 	github.com/coreos/go-systemd/v22 v22.5.0 // indirect | ||||
| 	github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc // indirect | ||||
| 	github.com/docker/go-units v0.5.0 // indirect | ||||
| 	github.com/godbus/dbus/v5 v5.1.0 // indirect | ||||
| 	github.com/moby/sys/userns v0.1.0 // indirect | ||||
| 	github.com/opencontainers/runtime-spec v1.2.0 // indirect | ||||
| 	github.com/pbnjay/memory v0.0.0-20210728143218-7b4eea64cf58 // indirect | ||||
| 	github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 // indirect | ||||
| 	github.com/rogpeppe/go-internal v1.13.1 // indirect | ||||
| 	github.com/sirupsen/logrus v1.9.3 // indirect | ||||
| 	go.uber.org/goleak v1.3.0 // indirect | ||||
| 	golang.org/x/exp v0.0.0-20241210194714-1829a127f884 // indirect | ||||
| 	golang.org/x/net v0.33.0 // indirect | ||||
| 	golang.org/x/sys v0.28.0 // indirect | ||||
| 	google.golang.org/genproto/googleapis/rpc v0.0.0-20241216192217-9240e9c98484 // indirect | ||||
| 	gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c // indirect | ||||
| ) | ||||
|   | ||||
							
								
								
									
										124
									
								
								go.sum
									
									
									
									
									
								
							
							
						
						
									
										124
									
								
								go.sum
									
									
									
									
									
								
							| @@ -1,33 +1,105 @@ | ||||
| dario.cat/mergo v1.0.0 h1:AGCNq9Evsj31mOgNPcLyXc+4PNABt905YmuqPYYpBWk= | ||||
| dario.cat/mergo v1.0.0/go.mod h1:uNxQE+84aUszobStD9th8a29P2fMDhsBdgRYvZOxGmk= | ||||
| github.com/DATA-DOG/go-sqlmock v1.5.0 h1:Shsta01QNfFxHCfpW6YH2STWB0MudeXXEWMr20OEh60= | ||||
| github.com/DATA-DOG/go-sqlmock v1.5.0/go.mod h1:f/Ixk793poVmq4qj/V1dPUg2JEAKC73Q5eFN3EC/SaM= | ||||
| github.com/golang/protobuf v1.5.0/go.mod h1:FsONVRAS9T7sI+LIUmWTfcYkHO4aIWwzhcaSAoJOfIk= | ||||
| github.com/golang/protobuf v1.5.3 h1:KhyjKVUg7Usr/dYsdSqoFveMYd5ko72D+zANwlG1mmg= | ||||
| github.com/golang/protobuf v1.5.3/go.mod h1:XVQd3VNwM+JqD3oG2Ue2ip4fOMUkwXdXDdiuN0vRsmY= | ||||
| github.com/google/go-cmp v0.5.5/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/gNBxE= | ||||
| github.com/google/go-cmp v0.5.9 h1:O2Tfq5qg4qc4AmwVlvv0oLiVAGB7enBSJ2x2DqQFi38= | ||||
| github.com/google/uuid v1.3.0 h1:t6JiXgmwXMjEs8VusXIJk2BXHsn+wx8BZdTaoZ5fu7I= | ||||
| github.com/google/uuid v1.3.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= | ||||
| dario.cat/mergo v1.0.1 h1:Ra4+bf83h2ztPIQYNP99R6m+Y7KfnARDfID+a+vLl4s= | ||||
| dario.cat/mergo v1.0.1/go.mod h1:uNxQE+84aUszobStD9th8a29P2fMDhsBdgRYvZOxGmk= | ||||
| github.com/DATA-DOG/go-sqlmock v1.5.2 h1:OcvFkGmslmlZibjAjaHm3L//6LiuBgolP7OputlJIzU= | ||||
| github.com/DATA-DOG/go-sqlmock v1.5.2/go.mod h1:88MAG/4G7SMwSE3CeA0ZKzrT5CiOU3OJ+JlNzwDqpNU= | ||||
| github.com/KimMachineGun/automemlimit v0.6.1 h1:ILa9j1onAAMadBsyyUJv5cack8Y1WT26yLj/V+ulKp8= | ||||
| github.com/KimMachineGun/automemlimit v0.6.1/go.mod h1:T7xYht7B8r6AG/AqFcUdc7fzd2bIdBKmepfP2S1svPY= | ||||
| github.com/ash3in/uuidv8 v1.2.0 h1:2oogGdtCPwaVtyvPPGin4TfZLtOGE5F+W++E880G6SI= | ||||
| github.com/ash3in/uuidv8 v1.2.0/go.mod h1:BnU0wJBxnzdEKmVg4xckBkD+VZuecTFTUP3M0dWgyY4= | ||||
| github.com/cilium/ebpf v0.16.0 h1:+BiEnHL6Z7lXnlGUsXQPPAE7+kenAd4ES8MQ5min0Ok= | ||||
| github.com/cilium/ebpf v0.16.0/go.mod h1:L7u2Blt2jMM/vLAVgjxluxtBKlz3/GWjB0dMOEngfwE= | ||||
| github.com/containerd/cgroups/v3 v3.0.4 h1:2fs7l3P0Qxb1nKWuJNFiwhp2CqiKzho71DQkDrHJIo4= | ||||
| github.com/containerd/cgroups/v3 v3.0.4/go.mod h1:SA5DLYnXO8pTGYiAHXz94qvLQTKfVM5GEVisn4jpins= | ||||
| github.com/containerd/log v0.1.0 h1:TCJt7ioM2cr/tfR8GPbGf9/VRAX8D2B4PjzCpfX540I= | ||||
| github.com/containerd/log v0.1.0/go.mod h1:VRRf09a7mHDIRezVKTRCrOq78v577GXq3bSa3EhrzVo= | ||||
| github.com/coreos/go-systemd/v22 v22.5.0 h1:RrqgGjYQKalulkV8NGVIfkXQf6YYmOyiJKk8iXXhfZs= | ||||
| github.com/coreos/go-systemd/v22 v22.5.0/go.mod h1:Y58oyj3AT4RCenI/lSvhwexgC+NSVTIJ3seZv2GcEnc= | ||||
| github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= | ||||
| github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= | ||||
| github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc h1:U9qPSI2PIWSS1VwoXQT9A3Wy9MM3WgvqSxFWenqJduM= | ||||
| github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38= | ||||
| github.com/docker/go-units v0.5.0 h1:69rxXcBk27SvSaaxTtLh/8llcHD8vYHT7WSdRZ/jvr4= | ||||
| github.com/docker/go-units v0.5.0/go.mod h1:fgPhTUdO+D/Jk86RDLlptpiXQzgHJF7gydDDbaIK4Dk= | ||||
| github.com/go-quicktest/qt v1.101.0 h1:O1K29Txy5P2OK0dGo59b7b0LR6wKfIhttaAhHUyn7eI= | ||||
| github.com/go-quicktest/qt v1.101.0/go.mod h1:14Bz/f7NwaXPtdYEgzsx46kqSxVwTbzVZsDC26tQJow= | ||||
| github.com/godbus/dbus/v5 v5.0.4/go.mod h1:xhWf0FNVPg57R7Z0UbKHbJfkEywrmjJnf7w5xrFpKfA= | ||||
| github.com/godbus/dbus/v5 v5.1.0 h1:4KLkAxT3aOY8Li4FRJe/KvhoNFFxo0m6fNuFUO8QJUk= | ||||
| github.com/godbus/dbus/v5 v5.1.0/go.mod h1:xhWf0FNVPg57R7Z0UbKHbJfkEywrmjJnf7w5xrFpKfA= | ||||
| github.com/golang/protobuf v1.5.4 h1:i7eJL8qZTpSEXOPTxNKhASYpMn+8e5Q6AdndVa1dWek= | ||||
| github.com/golang/protobuf v1.5.4/go.mod h1:lnTiLA8Wa4RWRcIUkrtSVa5nRhsEGBg48fD6rSs7xps= | ||||
| github.com/google/go-cmp v0.6.0 h1:ofyhxvXcZhMsU5ulbFiLKl/XBFqE1GSq7atu8tAmTRI= | ||||
| github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeNGIjoY= | ||||
| github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0= | ||||
| github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo= | ||||
| github.com/josharian/native v1.1.0 h1:uuaP0hAbW7Y4l0ZRQ6C9zfb7Mg1mbFKry/xzDAfmtLA= | ||||
| github.com/josharian/native v1.1.0/go.mod h1:7X/raswPFr05uY3HiLlYeyQntB6OO7E/d2Cu7qoaN2w= | ||||
| github.com/jsimonetti/rtnetlink/v2 v2.0.1 h1:xda7qaHDSVOsADNouv7ukSuicKZO7GgVUCXxpaIEIlM= | ||||
| github.com/jsimonetti/rtnetlink/v2 v2.0.1/go.mod h1:7MoNYNbb3UaDHtF8udiJo/RH6VsTKP1pqKLUTVCvToE= | ||||
| github.com/kisielk/sqlstruct v0.0.0-20201105191214-5f3e10d3ab46/go.mod h1:yyMNCyc/Ib3bDTKd379tNMpB/7/H5TjM2Y9QJ5THLbE= | ||||
| github.com/kr/pretty v0.2.1/go.mod h1:ipq/a2n7PKx3OHsz4KJII5eveXtPO4qwEXGdVfWzfnI= | ||||
| github.com/kr/pretty v0.3.1 h1:flRD4NNwYAUpkphVc1HcthR4KEIFJ65n8Mw5qdRn3LE= | ||||
| github.com/kr/pretty v0.3.1/go.mod h1:hoEshYVHaxMs3cyo3Yncou5ZscifuDolrwPKZanG3xk= | ||||
| github.com/kr/pty v1.1.1/go.mod h1:pFQYn66WHrOpPYNljwOMqo10TkYh1fy3cYio2l3bCsQ= | ||||
| github.com/kr/text v0.1.0/go.mod h1:4Jbv+DJW3UT/LiOwJeYQe1efqtUx/iVham/4vfdArNI= | ||||
| github.com/kr/text v0.2.0 h1:5Nx0Ya0ZqY2ygV366QzturHI13Jq95ApcVaJBhpS+AY= | ||||
| github.com/kr/text v0.2.0/go.mod h1:eLer722TekiGuMkidMxC/pM04lWEeraHUUmBw8l2grE= | ||||
| github.com/matoous/go-nanoid v1.5.1 h1:aCjdvTyO9LLnTIi0fgdXhOPPvOHjpXN6Ik9DaNjIct4= | ||||
| github.com/matoous/go-nanoid v1.5.1/go.mod h1:zyD2a71IubI24efhpvkJz+ZwfwagzgSO6UNiFsZKN7U= | ||||
| github.com/mdlayher/netlink v1.7.2 h1:/UtM3ofJap7Vl4QWCPDGXY8d3GIY2UGSDbK+QWmY8/g= | ||||
| github.com/mdlayher/netlink v1.7.2/go.mod h1:xraEF7uJbxLhc5fpHL4cPe221LI2bdttWlU+ZGLfQSw= | ||||
| github.com/mdlayher/socket v0.4.1 h1:eM9y2/jlbs1M615oshPQOHZzj6R6wMT7bX5NPiQvn2U= | ||||
| github.com/mdlayher/socket v0.4.1/go.mod h1:cAqeGjoufqdxWkD7DkpyS+wcefOtmu5OQ8KuoJGIReA= | ||||
| github.com/moby/sys/userns v0.1.0 h1:tVLXkFOxVu9A64/yh59slHVv9ahO9UIev4JZusOLG/g= | ||||
| github.com/moby/sys/userns v0.1.0/go.mod h1:IHUYgu/kao6N8YZlp9Cf444ySSvCmDlmzUcYfDHOl28= | ||||
| github.com/opencontainers/runtime-spec v1.2.0 h1:z97+pHb3uELt/yiAWD691HNHQIF07bE7dzrbT927iTk= | ||||
| github.com/opencontainers/runtime-spec v1.2.0/go.mod h1:jwyrGlmzljRJv/Fgzds9SsS/C5hL+LL3ko9hs6T5lQ0= | ||||
| github.com/patrickmn/go-cache v2.1.0+incompatible h1:HRMgzkcYKYpi3C8ajMPV8OFXaaRUnok+kx1WdO15EQc= | ||||
| github.com/patrickmn/go-cache v2.1.0+incompatible/go.mod h1:3Qf8kWWT7OJRJbdiICTKqZju1ZixQ/KpMGzzAfe6+WQ= | ||||
| github.com/pbnjay/memory v0.0.0-20210728143218-7b4eea64cf58 h1:onHthvaw9LFnH4t2DcNVpwGmV9E1BkGknEliJkfwQj0= | ||||
| github.com/pbnjay/memory v0.0.0-20210728143218-7b4eea64cf58/go.mod h1:DXv8WO4yhMYhSNPKjeNKa5WY9YCIEBRbNzFFPJbWO6Y= | ||||
| github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= | ||||
| github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2 h1:Jamvg5psRIccs7FGNTlIRMkT8wgtp5eCXdBlqhYGL6U= | ||||
| github.com/pmezard/go-difflib v1.0.1-0.20181226105442-5d4384ee4fb2/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4= | ||||
| github.com/prashantv/gostub v1.1.0 h1:BTyx3RfQjRHnUWaGF9oQos79AlQ5k8WNktv7VGvVH4g= | ||||
| github.com/prashantv/gostub v1.1.0/go.mod h1:A5zLQHz7ieHGG7is6LLXLz7I8+3LZzsrV0P1IAHhP5U= | ||||
| github.com/rogpeppe/go-internal v1.13.1 h1:KvO1DLK/DRN07sQ1LQKScxyZJuNnedQ5/wKSR38lUII= | ||||
| github.com/rogpeppe/go-internal v1.13.1/go.mod h1:uMEvuHeurkdAXX61udpOXGD/AzZDWNMNyH2VO9fmH0o= | ||||
| github.com/silas/dag v0.0.0-20220518035006-a7e85ada93c5 h1:G/FZtUu7a6NTWl3KUHMV9jkLAh/Rvtf03NWMHaEDl+E= | ||||
| github.com/silas/dag v0.0.0-20220518035006-a7e85ada93c5/go.mod h1:7RTUFBdIRC9nZ7/3RyRNH1bdqIShrDejd1YbLwgPS+I= | ||||
| golang.org/x/net v0.14.0 h1:BONx9s002vGdD9umnlX1Po8vOZmrgH34qlHcD1MfK14= | ||||
| golang.org/x/net v0.14.0/go.mod h1:PpSgVXXLK0OxS0F31C1/tv6XNguvCrnXIDrFMspZIUI= | ||||
| golang.org/x/sync v0.3.0 h1:ftCYgMx6zT/asHUrPw8BLLscYtGznsLAnjq5RH9P66E= | ||||
| golang.org/x/sync v0.3.0/go.mod h1:FU7BRWz2tNW+3quACPkgCx/L+uEAv1htQ0V83Z9Rj+Y= | ||||
| golang.org/x/sys v0.11.0 h1:eG7RXZHdqOJ1i+0lgLgCpSXAp6M3LYlAo6osgSi0xOM= | ||||
| golang.org/x/text v0.12.0 h1:k+n5B8goJNdU7hSvEtMUz3d1Q6D/XW4COJSJR6fN0mc= | ||||
| golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0= | ||||
| google.golang.org/genproto/googleapis/rpc v0.0.0-20230526203410-71b5a4ffd15e h1:NumxXLPfHSndr3wBBdeKiVHjGVFzi9RX2HwwQke94iY= | ||||
| google.golang.org/genproto/googleapis/rpc v0.0.0-20230526203410-71b5a4ffd15e/go.mod h1:66JfowdXAEgad5O9NnYcsNPLCPZJD++2L9X0PCMODrA= | ||||
| google.golang.org/grpc v1.57.0 h1:kfzNeI/klCGD2YPMUlaGNT3pxvYfga7smW3Vth8Zsiw= | ||||
| google.golang.org/grpc v1.57.0/go.mod h1:Sd+9RMTACXwmub0zcNY2c4arhtrbBYD1AUHI/dt16Mo= | ||||
| google.golang.org/protobuf v1.26.0-rc.1/go.mod h1:jlhhOSvTdKEhbULTjvd4ARK9grFBp09yW+WbY/TyQbw= | ||||
| google.golang.org/protobuf v1.26.0/go.mod h1:9q0QmTI4eRPtz6boOQmLYwt+qCgq0jsYwAQnmE0givc= | ||||
| google.golang.org/protobuf v1.31.0 h1:g0LDEJHgrBl9N9r17Ru3sqWhkIx2NB67okBHPwC7hs8= | ||||
| google.golang.org/protobuf v1.31.0/go.mod h1:HV8QOd/L58Z+nl8r43ehVNZIU/HEI6OcFqwMG9pJV4I= | ||||
| github.com/sirupsen/logrus v1.9.3 h1:dueUQJ1C2q9oE3F7wvmSGAaVtTmUizReu6fjN8uqzbQ= | ||||
| github.com/sirupsen/logrus v1.9.3/go.mod h1:naHLuLoDiP4jHNo9R0sCBMtWGeIprob74mVsIT4qYEQ= | ||||
| github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME= | ||||
| github.com/stretchr/testify v1.7.0/go.mod h1:6Fq8oRcR53rry900zMqJjRRixrwX3KX962/h/Wwjteg= | ||||
| github.com/stretchr/testify v1.10.0 h1:Xv5erBjTwe/5IxqUQTdXv5kgmIvbHo3QQyRwhJsOfJA= | ||||
| github.com/stretchr/testify v1.10.0/go.mod h1:r2ic/lqez/lEtzL7wO/rwa5dbSLXVDPFyf8C91i36aY= | ||||
| go.uber.org/automaxprocs v1.6.0 h1:O3y2/QNTOdbF+e/dpXNNW7Rx2hZ4sTIPyybbxyNqTUs= | ||||
| go.uber.org/automaxprocs v1.6.0/go.mod h1:ifeIMSnPZuznNm6jmdzmU3/bfk01Fe2fotchwEFJ8r8= | ||||
| go.uber.org/goleak v1.3.0 h1:2K3zAYmnTNqV73imy9J1T3WC+gmCePx2hEGkimedGto= | ||||
| go.uber.org/goleak v1.3.0/go.mod h1:CoHD4mav9JJNrW/WLlf7HGZPjdw8EucARQHekz1X6bE= | ||||
| go.unistack.org/micro-proto/v3 v3.4.1 h1:UTjLSRz2YZuaHk9iSlVqqsA50JQNAEK2ZFboGqtEa9Q= | ||||
| go.unistack.org/micro-proto/v3 v3.4.1/go.mod h1:okx/cnOhzuCX0ggl/vToatbCupi0O44diiiLLsZ93Zo= | ||||
| golang.org/x/exp v0.0.0-20241210194714-1829a127f884 h1:Y/Mj/94zIQQGHVSv1tTtQBDaQaJe62U9bkDZKKyhPCU= | ||||
| golang.org/x/exp v0.0.0-20241210194714-1829a127f884/go.mod h1:qj5a5QZpwLU2NLQudwIN5koi3beDhSAlJwa67PuM98c= | ||||
| golang.org/x/net v0.33.0 h1:74SYHlV8BIgHIFC/LrYkOGIwL19eTYXQ5wc6TBuO36I= | ||||
| golang.org/x/net v0.33.0/go.mod h1:HXLR5J+9DxmrqMwG9qjGCxZ+zKXxBru04zlTvWlWuN4= | ||||
| golang.org/x/sync v0.10.0 h1:3NQrjDixjgGwUOCaF8w2+VYHv0Ve/vGYSbdkTa98gmQ= | ||||
| golang.org/x/sync v0.10.0/go.mod h1:Czt+wKu1gCyEFDUtn0jG5QVvpJ6rzVqr5aXyt9drQfk= | ||||
| golang.org/x/sys v0.0.0-20220715151400-c0bba94af5f8/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg= | ||||
| golang.org/x/sys v0.28.0 h1:Fksou7UEQUWlKvIdsqzJmUmCX3cZuD2+P3XyyzwMhlA= | ||||
| golang.org/x/sys v0.28.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA= | ||||
| golang.org/x/text v0.21.0 h1:zyQAAkrwaneQ066sspRyJaG9VNi/YJ1NfzcGB3hZ/qo= | ||||
| golang.org/x/text v0.21.0/go.mod h1:4IBbMaMmOPCJ8SecivzSH54+73PCFmPWxNTLm+vZkEQ= | ||||
| golang.yandex/hasql/v2 v2.1.0 h1:7CaFFWeHoK5TvA+QvZzlKHlIN5sqNpqM8NSrXskZD/k= | ||||
| golang.yandex/hasql/v2 v2.1.0/go.mod h1:3Au1AxuJDCTXmS117BpbI6e+70kGWeyLR1qJAH6HdtA= | ||||
| google.golang.org/genproto/googleapis/rpc v0.0.0-20241216192217-9240e9c98484 h1:Z7FRVJPSMaHQxD0uXU8WdgFh8PseLM8Q8NzhnpMrBhQ= | ||||
| google.golang.org/genproto/googleapis/rpc v0.0.0-20241216192217-9240e9c98484/go.mod h1:lcTa1sDdWEIHMWlITnIczmw5w60CF9ffkb8Z+DVmmjA= | ||||
| google.golang.org/grpc v1.69.2 h1:U3S9QEtbXC0bYNvRtcoklF3xGtLViumSYxWykJS+7AU= | ||||
| google.golang.org/grpc v1.69.2/go.mod h1:vyjdE6jLBI76dgpDojsFGNaHlxdjXN9ghpnd2o7JGZ4= | ||||
| google.golang.org/protobuf v1.36.1 h1:yBPeRvTftaleIgM3PZ/WBIZ7XM/eEYAaEyCwvyjq/gk= | ||||
| google.golang.org/protobuf v1.36.1/go.mod h1:9fA7Ob0pmnwhb644+1+CVWFRbNajQ6iRojtC/QF5bRE= | ||||
| gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0= | ||||
| gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c h1:Hei/4ADfdWqJk1ZMxUNpqntNwaWcugrBjAiHlqqRiVk= | ||||
| gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c/go.mod h1:JHkPIbrfpd72SG/EVd6muEfDQjcINNoR0C8j2r3qZ4Q= | ||||
| gopkg.in/yaml.v3 v3.0.0-20200313102051-9f266ea9e77c/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM= | ||||
| gopkg.in/yaml.v3 v3.0.1 h1:fxVm/GzAzEWqLHuvctI91KS9hhNmmWOoWu0XTYJS7CA= | ||||
| gopkg.in/yaml.v3 v3.0.1/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM= | ||||
|   | ||||
							
								
								
									
										76
									
								
								hooks/metadata/metadata.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										76
									
								
								hooks/metadata/metadata.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,76 @@ | ||||
| package metadata | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/client" | ||||
| 	"go.unistack.org/micro/v3/metadata" | ||||
| 	"go.unistack.org/micro/v3/server" | ||||
| ) | ||||
|  | ||||
| var DefaultMetadataKeys = []string{"x-request-id"} | ||||
|  | ||||
| type hook struct { | ||||
| 	keys []string | ||||
| } | ||||
|  | ||||
| func NewHook(keys ...string) *hook { | ||||
| 	return &hook{keys: keys} | ||||
| } | ||||
|  | ||||
| func metadataCopy(ctx context.Context, keys []string) context.Context { | ||||
| 	if keys == nil { | ||||
| 		return ctx | ||||
| 	} | ||||
| 	if imd, iok := metadata.FromIncomingContext(ctx); iok && imd != nil { | ||||
| 		omd, ook := metadata.FromOutgoingContext(ctx) | ||||
| 		if !ook || omd == nil { | ||||
| 			omd = metadata.New(len(keys)) | ||||
| 		} | ||||
| 		for _, k := range keys { | ||||
| 			if v, ok := imd.Get(k); ok && v != "" { | ||||
| 				omd.Set(k, v) | ||||
| 			} | ||||
| 		} | ||||
| 		if !ook { | ||||
| 			ctx = metadata.NewOutgoingContext(ctx, omd) | ||||
| 		} | ||||
| 	} | ||||
| 	return ctx | ||||
| } | ||||
|  | ||||
| func (w *hook) ClientCall(next client.FuncCall) client.FuncCall { | ||||
| 	return func(ctx context.Context, req client.Request, rsp interface{}, opts ...client.CallOption) error { | ||||
| 		return next(metadataCopy(ctx, w.keys), req, rsp, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ClientStream(next client.FuncStream) client.FuncStream { | ||||
| 	return func(ctx context.Context, req client.Request, opts ...client.CallOption) (client.Stream, error) { | ||||
| 		return next(metadataCopy(ctx, w.keys), req, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ClientPublish(next client.FuncPublish) client.FuncPublish { | ||||
| 	return func(ctx context.Context, msg client.Message, opts ...client.PublishOption) error { | ||||
| 		return next(metadataCopy(ctx, w.keys), msg, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ClientBatchPublish(next client.FuncBatchPublish) client.FuncBatchPublish { | ||||
| 	return func(ctx context.Context, msgs []client.Message, opts ...client.PublishOption) error { | ||||
| 		return next(metadataCopy(ctx, w.keys), msgs, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ServerHandler(next server.FuncHandler) server.FuncHandler { | ||||
| 	return func(ctx context.Context, req server.Request, rsp interface{}) error { | ||||
| 		return next(metadataCopy(ctx, w.keys), req, rsp) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ServerSubscriber(next server.FuncSubHandler) server.FuncSubHandler { | ||||
| 	return func(ctx context.Context, msg server.Message) error { | ||||
| 		return next(metadataCopy(ctx, w.keys), msg) | ||||
| 	} | ||||
| } | ||||
							
								
								
									
										94
									
								
								hooks/recovery/recovery.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										94
									
								
								hooks/recovery/recovery.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,94 @@ | ||||
| package recovery | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"fmt" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/errors" | ||||
| 	"go.unistack.org/micro/v3/server" | ||||
| ) | ||||
|  | ||||
| func NewOptions(opts ...Option) Options { | ||||
| 	options := Options{ | ||||
| 		ServerHandlerFn:    DefaultServerHandlerFn, | ||||
| 		ServerSubscriberFn: DefaultServerSubscriberFn, | ||||
| 	} | ||||
| 	for _, o := range opts { | ||||
| 		o(&options) | ||||
| 	} | ||||
| 	return options | ||||
| } | ||||
|  | ||||
| type Options struct { | ||||
| 	ServerHandlerFn    func(context.Context, server.Request, interface{}, error) error | ||||
| 	ServerSubscriberFn func(context.Context, server.Message, error) error | ||||
| } | ||||
|  | ||||
| type Option func(*Options) | ||||
|  | ||||
| func ServerHandlerFunc(fn func(context.Context, server.Request, interface{}, error) error) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.ServerHandlerFn = fn | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func ServerSubscriberFunc(fn func(context.Context, server.Message, error) error) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.ServerSubscriberFn = fn | ||||
| 	} | ||||
| } | ||||
|  | ||||
| var ( | ||||
| 	DefaultServerHandlerFn = func(ctx context.Context, req server.Request, rsp interface{}, err error) error { | ||||
| 		return errors.BadRequest("", "%v", err) | ||||
| 	} | ||||
| 	DefaultServerSubscriberFn = func(ctx context.Context, req server.Message, err error) error { | ||||
| 		return errors.BadRequest("", "%v", err) | ||||
| 	} | ||||
| ) | ||||
|  | ||||
| var Hook = NewHook() | ||||
|  | ||||
| type hook struct { | ||||
| 	opts Options | ||||
| } | ||||
|  | ||||
| func NewHook(opts ...Option) *hook { | ||||
| 	return &hook{opts: NewOptions(opts...)} | ||||
| } | ||||
|  | ||||
| func (w *hook) ServerHandler(next server.FuncHandler) server.FuncHandler { | ||||
| 	return func(ctx context.Context, req server.Request, rsp interface{}) (err error) { | ||||
| 		defer func() { | ||||
| 			r := recover() | ||||
| 			switch verr := r.(type) { | ||||
| 			case nil: | ||||
| 				return | ||||
| 			case error: | ||||
| 				err = w.opts.ServerHandlerFn(ctx, req, rsp, verr) | ||||
| 			default: | ||||
| 				err = w.opts.ServerHandlerFn(ctx, req, rsp, fmt.Errorf("%v", r)) | ||||
| 			} | ||||
| 		}() | ||||
| 		err = next(ctx, req, rsp) | ||||
| 		return err | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ServerSubscriber(next server.FuncSubHandler) server.FuncSubHandler { | ||||
| 	return func(ctx context.Context, msg server.Message) (err error) { | ||||
| 		defer func() { | ||||
| 			r := recover() | ||||
| 			switch verr := r.(type) { | ||||
| 			case nil: | ||||
| 				return | ||||
| 			case error: | ||||
| 				err = w.opts.ServerSubscriberFn(ctx, msg, verr) | ||||
| 			default: | ||||
| 				err = w.opts.ServerSubscriberFn(ctx, msg, fmt.Errorf("%v", r)) | ||||
| 			} | ||||
| 		}() | ||||
| 		err = next(ctx, msg) | ||||
| 		return err | ||||
| 	} | ||||
| } | ||||
							
								
								
									
										139
									
								
								hooks/requestid/requestid.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										139
									
								
								hooks/requestid/requestid.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,139 @@ | ||||
| package requestid | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"net/textproto" | ||||
| 	 | ||||
| 	"go.unistack.org/micro/v3/client" | ||||
| 	"go.unistack.org/micro/v3/metadata" | ||||
| 	"go.unistack.org/micro/v3/server" | ||||
| 	"go.unistack.org/micro/v3/util/id" | ||||
| ) | ||||
|  | ||||
| type XRequestIDKey struct{} | ||||
|  | ||||
| // DefaultMetadataKey contains metadata key | ||||
| var DefaultMetadataKey = textproto.CanonicalMIMEHeaderKey("x-request-id") | ||||
|  | ||||
| // DefaultMetadataFunc wil be used if user not provide own func to fill metadata | ||||
| var DefaultMetadataFunc = func(ctx context.Context) (context.Context, error) { | ||||
| 	var xid string | ||||
|  | ||||
| 	cid, cok := ctx.Value(XRequestIDKey{}).(string) | ||||
| 	if cok && cid != "" { | ||||
| 		xid = cid | ||||
| 	} | ||||
|  | ||||
| 	imd, iok := metadata.FromIncomingContext(ctx) | ||||
| 	if !iok || imd == nil { | ||||
| 		imd = metadata.New(1) | ||||
| 		ctx = metadata.NewIncomingContext(ctx, imd) | ||||
| 	} | ||||
|  | ||||
| 	omd, ook := metadata.FromOutgoingContext(ctx) | ||||
| 	if !ook || omd == nil { | ||||
| 		omd = metadata.New(1) | ||||
| 		ctx = metadata.NewOutgoingContext(ctx, omd) | ||||
| 	} | ||||
|  | ||||
| 	if xid == "" { | ||||
| 		var id string | ||||
| 		if id, iok = imd.Get(DefaultMetadataKey); iok && id != "" { | ||||
| 			xid = id | ||||
| 		} | ||||
| 		if id, ook = omd.Get(DefaultMetadataKey); ook && id != "" { | ||||
| 			xid = id | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	if xid == "" { | ||||
| 		var err error | ||||
| 		xid, err = id.New() | ||||
| 		if err != nil { | ||||
| 			return ctx, err | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	if !cok { | ||||
| 		ctx = context.WithValue(ctx, XRequestIDKey{}, xid) | ||||
| 	} | ||||
|  | ||||
| 	if !iok { | ||||
| 		imd.Set(DefaultMetadataKey, xid) | ||||
| 	} | ||||
|  | ||||
| 	if !ook { | ||||
| 		omd.Set(DefaultMetadataKey, xid) | ||||
| 	} | ||||
|  | ||||
| 	return ctx, nil | ||||
| } | ||||
|  | ||||
| type hook struct{} | ||||
|  | ||||
| func NewHook() *hook { | ||||
| 	return &hook{} | ||||
| } | ||||
|  | ||||
| func (w *hook) ServerSubscriber(next server.FuncSubHandler) server.FuncSubHandler { | ||||
| 	return func(ctx context.Context, msg server.Message) error { | ||||
| 		var err error | ||||
| 		if xid, ok := msg.Header()[DefaultMetadataKey]; ok { | ||||
| 			ctx = context.WithValue(ctx, XRequestIDKey{}, xid) | ||||
| 		} | ||||
| 		if ctx, err = DefaultMetadataFunc(ctx); err != nil { | ||||
| 			return err | ||||
| 		} | ||||
| 		return next(ctx, msg) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ServerHandler(next server.FuncHandler) server.FuncHandler { | ||||
| 	return func(ctx context.Context, req server.Request, rsp interface{}) error { | ||||
| 		var err error | ||||
| 		if ctx, err = DefaultMetadataFunc(ctx); err != nil { | ||||
| 			return err | ||||
| 		} | ||||
| 		return next(ctx, req, rsp) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ClientBatchPublish(next client.FuncBatchPublish) client.FuncBatchPublish { | ||||
| 	return func(ctx context.Context, msgs []client.Message, opts ...client.PublishOption) error { | ||||
| 		var err error | ||||
| 		if ctx, err = DefaultMetadataFunc(ctx); err != nil { | ||||
| 			return err | ||||
| 		} | ||||
| 		return next(ctx, msgs, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ClientPublish(next client.FuncPublish) client.FuncPublish { | ||||
| 	return func(ctx context.Context, msg client.Message, opts ...client.PublishOption) error { | ||||
| 		var err error | ||||
| 		if ctx, err = DefaultMetadataFunc(ctx); err != nil { | ||||
| 			return err | ||||
| 		} | ||||
| 		return next(ctx, msg, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ClientCall(next client.FuncCall) client.FuncCall { | ||||
| 	return func(ctx context.Context, req client.Request, rsp interface{}, opts ...client.CallOption) error { | ||||
| 		var err error | ||||
| 		if ctx, err = DefaultMetadataFunc(ctx); err != nil { | ||||
| 			return err | ||||
| 		} | ||||
| 		return next(ctx, req, rsp, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ClientStream(next client.FuncStream) client.FuncStream { | ||||
| 	return func(ctx context.Context, req client.Request, opts ...client.CallOption) (client.Stream, error) { | ||||
| 		var err error | ||||
| 		if ctx, err = DefaultMetadataFunc(ctx); err != nil { | ||||
| 			return nil, err | ||||
| 		} | ||||
| 		return next(ctx, req, opts...) | ||||
| 	} | ||||
| } | ||||
							
								
								
									
										33
									
								
								hooks/requestid/requestid_test.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										33
									
								
								hooks/requestid/requestid_test.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,33 @@ | ||||
| package requestid | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"testing" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/metadata" | ||||
| ) | ||||
|  | ||||
| func TestDefaultMetadataFunc(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
|  | ||||
| 	nctx, err := DefaultMetadataFunc(ctx) | ||||
| 	if err != nil { | ||||
| 		t.Fatalf("%v", err) | ||||
| 	} | ||||
|  | ||||
| 	imd, ok := metadata.FromIncomingContext(nctx) | ||||
| 	if !ok { | ||||
| 		t.Fatalf("md missing in incoming context") | ||||
| 	} | ||||
| 	omd, ok := metadata.FromOutgoingContext(nctx) | ||||
| 	if !ok { | ||||
| 		t.Fatalf("md missing in outgoing context") | ||||
| 	} | ||||
|  | ||||
| 	_, iok := imd.Get(DefaultMetadataKey) | ||||
| 	_, ook := omd.Get(DefaultMetadataKey) | ||||
|  | ||||
| 	if !iok || !ook { | ||||
| 		t.Fatalf("missing metadata key value") | ||||
| 	} | ||||
| } | ||||
							
								
								
									
										51
									
								
								hooks/sql/common.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										51
									
								
								hooks/sql/common.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,51 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"database/sql/driver" | ||||
| 	"errors" | ||||
| 	"fmt" | ||||
| 	"runtime" | ||||
| ) | ||||
|  | ||||
| //go:generate sh -c "go run gen.go > wrap_gen.go" | ||||
|  | ||||
| // namedValueToValue converts driver arguments of NamedValue format to Value format. Implemented in the same way as in | ||||
| // database/sql ctxutil.go. | ||||
| func namedValueToValue(named []driver.NamedValue) ([]driver.Value, error) { | ||||
| 	dargs := make([]driver.Value, len(named)) | ||||
| 	for n, param := range named { | ||||
| 		if len(param.Name) > 0 { | ||||
| 			return nil, errors.New("sql: driver does not support the use of Named Parameters") | ||||
| 		} | ||||
| 		dargs[n] = param.Value | ||||
| 	} | ||||
| 	return dargs, nil | ||||
| } | ||||
|  | ||||
| // namedValueToLabels convert driver arguments to interface{} slice | ||||
| func namedValueToLabels(named []driver.NamedValue) []interface{} { | ||||
| 	largs := make([]interface{}, 0, len(named)*2) | ||||
| 	var name string | ||||
| 	for _, param := range named { | ||||
| 		if param.Name != "" { | ||||
| 			name = param.Name | ||||
| 		} else { | ||||
| 			name = fmt.Sprintf("$%d", param.Ordinal) | ||||
| 		} | ||||
| 		largs = append(largs, fmt.Sprintf("%s=%v", name, param.Value)) | ||||
| 	} | ||||
| 	return largs | ||||
| } | ||||
|  | ||||
| // getCallerName get the name of the function A where A() -> B() -> GetFunctionCallerName() | ||||
| func getCallerName() string { | ||||
| 	pc, _, _, ok := runtime.Caller(3) | ||||
| 	details := runtime.FuncForPC(pc) | ||||
| 	var callerName string | ||||
| 	if ok && details != nil { | ||||
| 		callerName = details.Name() | ||||
| 	} else { | ||||
| 		callerName = labelUnknown | ||||
| 	} | ||||
| 	return callerName | ||||
| } | ||||
							
								
								
									
										467
									
								
								hooks/sql/conn.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										467
									
								
								hooks/sql/conn.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,467 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"database/sql/driver" | ||||
| 	"fmt" | ||||
| 	"time" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/hooks/requestid" | ||||
| 	"go.unistack.org/micro/v3/tracer" | ||||
| ) | ||||
|  | ||||
| var ( | ||||
| 	_ driver.Conn               = (*wrapperConn)(nil) | ||||
| 	_ driver.ConnBeginTx        = (*wrapperConn)(nil) | ||||
| 	_ driver.ConnPrepareContext = (*wrapperConn)(nil) | ||||
| 	_ driver.Pinger             = (*wrapperConn)(nil) | ||||
| 	_ driver.Validator          = (*wrapperConn)(nil) | ||||
| 	_ driver.Queryer            = (*wrapperConn)(nil) // nolint:staticcheck | ||||
| 	_ driver.QueryerContext     = (*wrapperConn)(nil) | ||||
| 	_ driver.Execer             = (*wrapperConn)(nil) // nolint:staticcheck | ||||
| 	_ driver.ExecerContext      = (*wrapperConn)(nil) | ||||
| 	//	_ driver.Connector | ||||
| 	//	_ driver.Driver | ||||
| 	//	_ driver.DriverContext | ||||
| ) | ||||
|  | ||||
| // wrapperConn defines a wrapper for driver.Conn | ||||
| type wrapperConn struct { | ||||
| 	d     *wrapperDriver | ||||
| 	dname string | ||||
| 	conn  driver.Conn | ||||
| 	opts  Options | ||||
| 	ctx   context.Context | ||||
| 	//span  tracer.Span | ||||
| } | ||||
|  | ||||
| // Close implements driver.Conn Close | ||||
| func (w *wrapperConn) Close() error { | ||||
| 	var ctx context.Context | ||||
| 	if w.ctx != nil { | ||||
| 		ctx = w.ctx | ||||
| 	} else { | ||||
| 		ctx = context.Background() | ||||
| 	} | ||||
| 	_ = ctx | ||||
| 	labels := []string{labelMethod, "Close"} | ||||
| 	ts := time.Now() | ||||
| 	err := w.conn.Close() | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 	} else { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	} | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "Close", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| // Begin implements driver.Conn Begin | ||||
| func (w *wrapperConn) Begin() (driver.Tx, error) { | ||||
| 	var ctx context.Context | ||||
| 	if w.ctx != nil { | ||||
| 		ctx = w.ctx | ||||
| 	} else { | ||||
| 		ctx = context.Background() | ||||
| 	} | ||||
|  | ||||
| 	labels := []string{labelMethod, "Begin"} | ||||
| 	ts := time.Now() | ||||
| 	tx, err := w.conn.Begin() // nolint:staticcheck | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 		w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 		w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 		/* | ||||
| 			if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 				w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "Begin", getCallerName(), td, err)...) | ||||
| 			} | ||||
| 		*/ | ||||
| 		return nil, err | ||||
| 	} | ||||
| 	w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "Begin", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return &wrapperTx{tx: tx, opts: w.opts, ctx: ctx}, nil | ||||
| } | ||||
|  | ||||
| // BeginTx implements driver.ConnBeginTx BeginTx | ||||
| func (w *wrapperConn) BeginTx(ctx context.Context, opts driver.TxOptions) (driver.Tx, error) { | ||||
| 	name := getQueryName(ctx) | ||||
| 	nctx, span := w.opts.Tracer.Start(ctx, "sdk.database", tracer.WithSpanKind(tracer.SpanKindClient)) | ||||
| 	span.AddLabels("db.method", "BeginTx") | ||||
| 	span.AddLabels("db.statement", name) | ||||
| 	if id, ok := ctx.Value(requestid.XRequestIDKey{}).(string); ok { | ||||
| 		span.AddLabels("x-request-id", id) | ||||
| 	} | ||||
| 	labels := []string{labelMethod, "BeginTx", labelQuery, name} | ||||
|  | ||||
| 	connBeginTx, ok := w.conn.(driver.ConnBeginTx) | ||||
| 	if !ok { | ||||
| 		return w.Begin() | ||||
| 	} | ||||
|  | ||||
| 	ts := time.Now() | ||||
| 	tx, err := connBeginTx.BeginTx(nctx, opts) | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 		w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 		w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 		span.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 		/* | ||||
| 			if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 				w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "BeginTx", getCallerName(), td, err)...) | ||||
| 			} | ||||
| 		*/ | ||||
| 		return nil, err | ||||
| 	} | ||||
| 	w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "BeginTx", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return &wrapperTx{tx: tx, opts: w.opts, ctx: ctx, span: span}, nil | ||||
| } | ||||
|  | ||||
| // Prepare implements driver.Conn Prepare | ||||
| func (w *wrapperConn) Prepare(query string) (driver.Stmt, error) { | ||||
| 	var ctx context.Context | ||||
| 	if w.ctx != nil { | ||||
| 		ctx = w.ctx | ||||
| 	} else { | ||||
| 		ctx = context.Background() | ||||
| 	} | ||||
| 	_ = ctx | ||||
| 	labels := []string{labelMethod, "Prepare", labelQuery, getCallerName()} | ||||
| 	ts := time.Now() | ||||
| 	stmt, err := w.conn.Prepare(query) | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 		w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 		w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 		/* | ||||
| 			if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 				w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "Prepare", getCallerName(), td, err)...) | ||||
| 			} | ||||
| 		*/ | ||||
| 		return nil, err | ||||
| 	} | ||||
| 	w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "Prepare", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return wrapStmt(stmt, query, w.opts), nil | ||||
| } | ||||
|  | ||||
| // PrepareContext implements driver.ConnPrepareContext PrepareContext | ||||
| func (w *wrapperConn) PrepareContext(ctx context.Context, query string) (driver.Stmt, error) { | ||||
| 	var nctx context.Context | ||||
| 	var span tracer.Span | ||||
|  | ||||
| 	name := getQueryName(ctx) | ||||
| 	if w.ctx != nil { | ||||
| 		nctx, span = w.opts.Tracer.Start(w.ctx, "sdk.database", tracer.WithSpanKind(tracer.SpanKindClient)) | ||||
| 	} else { | ||||
| 		nctx, span = w.opts.Tracer.Start(ctx, "sdk.database", tracer.WithSpanKind(tracer.SpanKindClient)) | ||||
| 	} | ||||
| 	span.AddLabels("db.method", "PrepareContext") | ||||
| 	span.AddLabels("db.statement", name) | ||||
| 	if id, ok := ctx.Value(requestid.XRequestIDKey{}).(string); ok { | ||||
| 		span.AddLabels("x-request-id", id) | ||||
| 	} | ||||
| 	labels := []string{labelMethod, "PrepareContext", labelQuery, name} | ||||
| 	conn, ok := w.conn.(driver.ConnPrepareContext) | ||||
| 	if !ok { | ||||
| 		return w.Prepare(query) | ||||
| 	} | ||||
|  | ||||
| 	ts := time.Now() | ||||
| 	stmt, err := conn.PrepareContext(nctx, query) | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 		w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 		w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 		span.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 		/* | ||||
| 			if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 				w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "PrepareContext", getCallerName(), td, err)...) | ||||
| 			} | ||||
| 		*/ | ||||
| 		return nil, err | ||||
| 	} | ||||
| 	w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "PrepareContext", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return wrapStmt(stmt, query, w.opts), nil | ||||
| } | ||||
|  | ||||
| // Exec implements driver.Execer Exec | ||||
| func (w *wrapperConn) Exec(query string, args []driver.Value) (driver.Result, error) { | ||||
| 	var ctx context.Context | ||||
| 	if w.ctx != nil { | ||||
| 		ctx = w.ctx | ||||
| 	} else { | ||||
| 		ctx = context.Background() | ||||
| 	} | ||||
| 	_ = ctx | ||||
| 	labels := []string{labelMethod, "Exec", labelQuery, getCallerName()} | ||||
|  | ||||
| 	// nolint:staticcheck | ||||
| 	conn, ok := w.conn.(driver.Execer) | ||||
| 	if !ok { | ||||
| 		return nil, driver.ErrSkip | ||||
| 	} | ||||
|  | ||||
| 	ts := time.Now() | ||||
| 	res, err := conn.Exec(query, args) | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 	} else { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	} | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "Exec", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return res, err | ||||
| } | ||||
|  | ||||
| // Exec implements driver.StmtExecContext ExecContext | ||||
| func (w *wrapperConn) ExecContext(ctx context.Context, query string, args []driver.NamedValue) (driver.Result, error) { | ||||
| 	var nctx context.Context | ||||
| 	var span tracer.Span | ||||
|  | ||||
| 	name := getQueryName(ctx) | ||||
| 	if w.ctx != nil { | ||||
| 		nctx, span = w.opts.Tracer.Start(w.ctx, "sdk.database", tracer.WithSpanKind(tracer.SpanKindClient)) | ||||
| 	} else { | ||||
| 		nctx, span = w.opts.Tracer.Start(ctx, "sdk.database", tracer.WithSpanKind(tracer.SpanKindClient)) | ||||
| 	} | ||||
| 	span.AddLabels("db.method", "ExecContext") | ||||
| 	span.AddLabels("db.statement", name) | ||||
| 	if id, ok := ctx.Value(requestid.XRequestIDKey{}).(string); ok { | ||||
| 		span.AddLabels("x-request-id", id) | ||||
| 	} | ||||
| 	defer span.Finish() | ||||
| 	if len(args) > 0 { | ||||
| 		span.AddLabels("db.args", fmt.Sprintf("%v", namedValueToLabels(args))) | ||||
| 	} | ||||
| 	labels := []string{labelMethod, "ExecContext", labelQuery, name} | ||||
|  | ||||
| 	conn, ok := w.conn.(driver.ExecerContext) | ||||
| 	if !ok { | ||||
| 		// nolint:staticcheck | ||||
| 		return nil, driver.ErrSkip | ||||
| 	} | ||||
|  | ||||
| 	ts := time.Now() | ||||
| 	res, err := conn.ExecContext(nctx, query, args) | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 		span.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 	} else { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	} | ||||
| 	w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "ExecContext", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return res, err | ||||
| } | ||||
|  | ||||
| // Ping implements driver.Pinger Ping | ||||
| func (w *wrapperConn) Ping(ctx context.Context) error { | ||||
| 	conn, ok := w.conn.(driver.Pinger) | ||||
|  | ||||
| 	if !ok { | ||||
| 		// fallback path to check db alive | ||||
| 		pc, err := w.d.Open(w.dname) | ||||
| 		if err != nil { | ||||
| 			return err | ||||
| 		} | ||||
| 		return pc.Close() | ||||
| 	} | ||||
|  | ||||
| 	var nctx context.Context //nolint: gosimple | ||||
| 	nctx = ctx | ||||
| 	/* | ||||
| 		var span tracer.Span | ||||
| 		if w.ctx != nil { | ||||
| 			nctx, span = w.opts.Tracer.Start(w.ctx, "sdk.database", tracer.WithSpanKind(tracer.SpanKindClient)) | ||||
| 		} else { | ||||
| 			nctx, span = w.opts.Tracer.Start(ctx, "sdk.database", tracer.WithSpanKind(tracer.SpanKindClient)) | ||||
| 		} | ||||
| 		span.AddLabels("db.method", "Ping") | ||||
| 		defer span.Finish() | ||||
| 	*/ | ||||
| 	labels := []string{labelMethod, "Ping"} | ||||
| 	ts := time.Now() | ||||
| 	err := conn.Ping(nctx) | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 		// span.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 		/* | ||||
| 			if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 				w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "Ping", getCallerName(), td, err)...) | ||||
| 			} | ||||
| 		*/ | ||||
| 		return err | ||||
| 	} else { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	} | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
|  | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| // Query implements driver.Queryer Query | ||||
| func (w *wrapperConn) Query(query string, args []driver.Value) (driver.Rows, error) { | ||||
| 	var ctx context.Context | ||||
| 	if w.ctx != nil { | ||||
| 		ctx = w.ctx | ||||
| 	} else { | ||||
| 		ctx = context.Background() | ||||
| 	} | ||||
| 	_ = ctx | ||||
| 	// nolint:staticcheck | ||||
| 	conn, ok := w.conn.(driver.Queryer) | ||||
| 	if !ok { | ||||
| 		return nil, driver.ErrSkip | ||||
| 	} | ||||
|  | ||||
| 	labels := []string{labelMethod, "Query", labelQuery, getCallerName()} | ||||
| 	ts := time.Now() | ||||
| 	rows, err := conn.Query(query, args) | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 	} else { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	} | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "Query", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return rows, err | ||||
| } | ||||
|  | ||||
| // QueryContext implements Driver.QueryerContext QueryContext | ||||
| func (w *wrapperConn) QueryContext(ctx context.Context, query string, args []driver.NamedValue) (driver.Rows, error) { | ||||
| 	var nctx context.Context | ||||
| 	var span tracer.Span | ||||
|  | ||||
| 	name := getQueryName(ctx) | ||||
| 	if w.ctx != nil { | ||||
| 		nctx, span = w.opts.Tracer.Start(w.ctx, "sdk.database", tracer.WithSpanKind(tracer.SpanKindClient)) | ||||
| 	} else { | ||||
| 		nctx, span = w.opts.Tracer.Start(ctx, "sdk.database", tracer.WithSpanKind(tracer.SpanKindClient)) | ||||
| 	} | ||||
| 	span.AddLabels("db.method", "QueryContext") | ||||
| 	span.AddLabels("db.statement", name) | ||||
| 	if id, ok := ctx.Value(requestid.XRequestIDKey{}).(string); ok { | ||||
| 		span.AddLabels("x-request-id", id) | ||||
| 	} | ||||
| 	defer span.Finish() | ||||
| 	if len(args) > 0 { | ||||
| 		span.AddLabels("db.args", fmt.Sprintf("%v", namedValueToLabels(args))) | ||||
| 	} | ||||
| 	labels := []string{labelMethod, "QueryContext", labelQuery, name} | ||||
| 	conn, ok := w.conn.(driver.QueryerContext) | ||||
| 	if !ok { | ||||
| 		return nil, driver.ErrSkip | ||||
| 	} | ||||
|  | ||||
| 	ts := time.Now() | ||||
| 	rows, err := conn.QueryContext(nctx, query, args) | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 		span.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 	} else { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	} | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "QueryContext", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return rows, err | ||||
| } | ||||
|  | ||||
| // CheckNamedValue implements driver.NamedValueChecker | ||||
| func (w *wrapperConn) CheckNamedValue(v *driver.NamedValue) error { | ||||
| 	s, ok := w.conn.(driver.NamedValueChecker) | ||||
| 	if !ok { | ||||
| 		return driver.ErrSkip | ||||
| 	} | ||||
| 	return s.CheckNamedValue(v) | ||||
| } | ||||
|  | ||||
| // IsValid implements driver.Validator | ||||
| func (w *wrapperConn) IsValid() bool { | ||||
| 	v, ok := w.conn.(driver.Validator) | ||||
| 	if !ok { | ||||
| 		return w.conn != nil | ||||
| 	} | ||||
| 	return v.IsValid() | ||||
| } | ||||
|  | ||||
| func (w *wrapperConn) ResetSession(ctx context.Context) error { | ||||
| 	s, ok := w.conn.(driver.SessionResetter) | ||||
| 	if !ok { | ||||
| 		return driver.ErrSkip | ||||
| 	} | ||||
| 	return s.ResetSession(ctx) | ||||
| } | ||||
							
								
								
									
										94
									
								
								hooks/sql/driver.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										94
									
								
								hooks/sql/driver.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,94 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"database/sql/driver" | ||||
| 	"time" | ||||
| ) | ||||
|  | ||||
| var ( | ||||
| // _ driver.DriverContext = (*wrapperDriver)(nil) | ||||
| // _ driver.Connector     = (*wrapperDriver)(nil) | ||||
| ) | ||||
|  | ||||
| /* | ||||
| type conn interface { | ||||
| 	driver.Pinger | ||||
| 	driver.Execer | ||||
| 	driver.ExecerContext | ||||
| 	driver.Queryer | ||||
| 	driver.QueryerContext | ||||
| 	driver.Conn | ||||
| 	driver.ConnPrepareContext | ||||
| 	driver.ConnBeginTx | ||||
| } | ||||
| */ | ||||
|  | ||||
| // wrapperDriver defines a wrapper for driver.Driver | ||||
| type wrapperDriver struct { | ||||
| 	driver driver.Driver | ||||
| 	opts   Options | ||||
| 	ctx    context.Context | ||||
| } | ||||
|  | ||||
| // NewWrapper creates and returns a new SQL driver with passed capabilities | ||||
| func NewWrapper(d driver.Driver, opts ...Option) driver.Driver { | ||||
| 	return &wrapperDriver{driver: d, opts: NewOptions(opts...), ctx: context.Background()} | ||||
| } | ||||
|  | ||||
| type wrappedConnector struct { | ||||
| 	connector driver.Connector | ||||
| //	name      string | ||||
| 	opts      Options | ||||
| 	ctx       context.Context | ||||
| } | ||||
|  | ||||
| func NewWrapperConnector(c driver.Connector, opts ...Option) driver.Connector { | ||||
| 	return &wrappedConnector{connector: c, opts: NewOptions(opts...), ctx: context.Background()} | ||||
| } | ||||
|  | ||||
| // Connect implements driver.Driver Connect | ||||
| func (w *wrappedConnector) Connect(ctx context.Context) (driver.Conn, error) { | ||||
| 	return w.connector.Connect(ctx) | ||||
| } | ||||
|  | ||||
| // Driver implements driver.Driver Driver | ||||
| func (w *wrappedConnector) Driver() driver.Driver { | ||||
| 	return w.connector.Driver() | ||||
| } | ||||
|  | ||||
| /* | ||||
| // Connect implements driver.Driver OpenConnector | ||||
| func (w *wrapperDriver) OpenConnector(name string) (driver.Conn, error) { | ||||
| 	return &wrapperConnector{driver: w.driver, name: name, opts: w.opts}, nil | ||||
| } | ||||
| */ | ||||
|  | ||||
| // Open implements driver.Driver Open | ||||
| func (w *wrapperDriver) Open(name string) (driver.Conn, error) { | ||||
| 	// ctx, cancel := context.WithTimeout(context.Background(), 60*time.Second) // Ensure eventual timeout | ||||
| 	// defer cancel() | ||||
|  | ||||
| 	/* | ||||
| 		connector, err := w.OpenConnector(name) | ||||
| 		if err != nil { | ||||
| 			return nil, err | ||||
| 		} | ||||
| 		return connector.Connect(ctx) | ||||
| 	*/ | ||||
|  | ||||
| 	ts := time.Now() | ||||
| 	c, err := w.driver.Open(name) | ||||
| 	td := time.Since(ts) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled { | ||||
| 			w.opts.Logger.Log(w.ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(w.ctx, "Open", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	_ = td | ||||
| 	if err != nil { | ||||
| 		return nil, err | ||||
| 	} | ||||
|  | ||||
| 	return wrapConn(c, w.opts), nil | ||||
| } | ||||
							
								
								
									
										165
									
								
								hooks/sql/gen.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										165
									
								
								hooks/sql/gen.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,165 @@ | ||||
| //go:build ignore | ||||
|  | ||||
| package main | ||||
|  | ||||
| import ( | ||||
| 	"bytes" | ||||
| 	"crypto/md5" | ||||
| 	"fmt" | ||||
| 	"io" | ||||
| 	"sort" | ||||
| 	"strings" | ||||
| ) | ||||
|  | ||||
| var connIfaces = []string{ | ||||
| 	"driver.ConnBeginTx", | ||||
| 	"driver.ConnPrepareContext", | ||||
| 	"driver.Execer", | ||||
| 	"driver.ExecerContext", | ||||
| 	"driver.NamedValueChecker", | ||||
| 	"driver.Pinger", | ||||
| 	"driver.Queryer", | ||||
| 	"driver.QueryerContext", | ||||
| 	"driver.SessionResetter", | ||||
| 	"driver.Validator", | ||||
| } | ||||
|  | ||||
| var stmtIfaces = []string{ | ||||
| 	"driver.StmtExecContext", | ||||
| 	"driver.StmtQueryContext", | ||||
| 	"driver.ColumnConverter", | ||||
| 	"driver.NamedValueChecker", | ||||
| } | ||||
|  | ||||
| func getHash(s []string) string { | ||||
| 	h := md5.New() | ||||
| 	io.WriteString(h, strings.Join(s, "|")) | ||||
| 	return fmt.Sprintf("%x", h.Sum(nil)) | ||||
| } | ||||
|  | ||||
| func main() { | ||||
| 	comboConn := all(connIfaces) | ||||
|  | ||||
| 	sort.Slice(comboConn, func(i, j int) bool { | ||||
| 		return len(comboConn[i]) < len(comboConn[j]) | ||||
| 	}) | ||||
|  | ||||
| 	comboStmt := all(stmtIfaces) | ||||
|  | ||||
| 	sort.Slice(comboStmt, func(i, j int) bool { | ||||
| 		return len(comboStmt[i]) < len(comboStmt[j]) | ||||
| 	}) | ||||
|  | ||||
| 	b := bytes.NewBuffer(nil) | ||||
| 	b.WriteString("// Code generated. DO NOT EDIT.\n\n") | ||||
| 	b.WriteString("package sql\n\n") | ||||
| 	b.WriteString(`import "database/sql/driver"`) | ||||
| 	b.WriteString("\n\n") | ||||
|  | ||||
| 	b.WriteString("func wrapConn(dc driver.Conn, opts Options) driver.Conn {\n") | ||||
| 	b.WriteString("\tc := &wrapperConn{conn: dc, opts: opts}\n") | ||||
|  | ||||
| 	for idx := len(comboConn) - 1; idx >= 0; idx-- { | ||||
| 		ifaces := comboConn[idx] | ||||
| 		n := len(ifaces) | ||||
| 		if n == 0 { | ||||
| 			continue | ||||
| 		} | ||||
| 		h := getHash(ifaces) | ||||
| 		b.WriteString(fmt.Sprintf("\tif _, ok := dc.(wrapConn%04d_%s); ok {\n", n, h)) | ||||
| 		b.WriteString("\t\treturn struct {\n") | ||||
| 		b.WriteString(strings.Join(append([]string{"\t\t\tdriver.Conn"}, ifaces...), "\n\t\t\t")) | ||||
| 		b.WriteString("\n\t\t}{") | ||||
| 		for idx := range ifaces { | ||||
| 			if idx > 0 { | ||||
| 				b.WriteString(", ") | ||||
| 				b.WriteString("c") | ||||
| 			} else if idx == 0 { | ||||
| 				b.WriteString("c") | ||||
| 			} else { | ||||
| 				b.WriteString("c") | ||||
| 			} | ||||
| 		} | ||||
| 		b.WriteString(", c}\n") | ||||
| 		b.WriteString("\t}\n\n") | ||||
| 	} | ||||
| 	b.WriteString("\treturn c\n") | ||||
| 	b.WriteString("}\n\n") | ||||
|  | ||||
| 	for idx := len(comboConn) - 1; idx >= 0; idx-- { | ||||
| 		ifaces := comboConn[idx] | ||||
| 		n := len(ifaces) | ||||
| 		if n == 0 { | ||||
| 			continue | ||||
| 		} | ||||
| 		h := getHash(ifaces) | ||||
| 		b.WriteString(fmt.Sprintf("// %s\n", strings.Join(ifaces, "|"))) | ||||
| 		b.WriteString(fmt.Sprintf("type wrapConn%04d_%s interface {\n", n, h)) | ||||
| 		for _, iface := range ifaces { | ||||
| 			b.WriteString(fmt.Sprintf("\t%s\n", iface)) | ||||
| 		} | ||||
| 		b.WriteString("}\n\n") | ||||
| 	} | ||||
|  | ||||
| 	b.WriteString("func wrapStmt(stmt driver.Stmt, query string, opts Options) driver.Stmt {\n") | ||||
| 	b.WriteString("\tc := &wrapperStmt{stmt: stmt, query: query, opts: opts}\n") | ||||
|  | ||||
| 	for idx := len(comboStmt) - 1; idx >= 0; idx-- { | ||||
| 		ifaces := comboStmt[idx] | ||||
| 		n := len(ifaces) | ||||
| 		if n == 0 { | ||||
| 			continue | ||||
| 		} | ||||
| 		h := getHash(ifaces) | ||||
| 		b.WriteString(fmt.Sprintf("\tif _, ok := stmt.(wrapStmt%04d_%s); ok {\n", n, h)) | ||||
| 		b.WriteString("\t\treturn struct {\n") | ||||
| 		b.WriteString(strings.Join(append([]string{"\t\t\tdriver.Stmt"}, ifaces...), "\n\t\t\t")) | ||||
| 		b.WriteString("\n\t\t}{") | ||||
| 		for idx := range ifaces { | ||||
| 			if idx > 0 { | ||||
| 				b.WriteString(", ") | ||||
| 				b.WriteString("c") | ||||
| 			} else if idx == 0 { | ||||
| 				b.WriteString("c") | ||||
| 			} else { | ||||
| 				b.WriteString("c") | ||||
| 			} | ||||
| 		} | ||||
| 		b.WriteString(", c}\n") | ||||
| 		b.WriteString("\t}\n\n") | ||||
| 	} | ||||
| 	b.WriteString("\treturn c\n") | ||||
| 	b.WriteString("}\n") | ||||
|  | ||||
| 	for idx := len(comboStmt) - 1; idx >= 0; idx-- { | ||||
| 		ifaces := comboStmt[idx] | ||||
| 		n := len(ifaces) | ||||
| 		if n == 0 { | ||||
| 			continue | ||||
| 		} | ||||
| 		h := getHash(ifaces) | ||||
| 		b.WriteString(fmt.Sprintf("\n// %s\n", strings.Join(ifaces, "|"))) | ||||
| 		b.WriteString(fmt.Sprintf("type wrapStmt%04d_%s interface {\n", n, h)) | ||||
| 		for _, iface := range ifaces { | ||||
| 			b.WriteString(fmt.Sprintf("\t%s\n", iface)) | ||||
| 		} | ||||
| 		b.WriteString("}\n") | ||||
| 	} | ||||
|  | ||||
| 	fmt.Printf("%s", b.String()) | ||||
| } | ||||
|  | ||||
| // all returns all combinations for a given string array. | ||||
| func all[T any](set []T) (subsets [][]T) { | ||||
| 	length := uint(len(set)) | ||||
| 	for subsetBits := 1; subsetBits < (1 << length); subsetBits++ { | ||||
| 		var subset []T | ||||
| 		for object := uint(0); object < length; object++ { | ||||
| 			if (subsetBits>>object)&1 == 1 { | ||||
| 				subset = append(subset, set[object]) | ||||
| 			} | ||||
| 		} | ||||
| 		subsets = append(subsets, subset) | ||||
| 	} | ||||
| 	return subsets | ||||
| } | ||||
							
								
								
									
										172
									
								
								hooks/sql/options.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										172
									
								
								hooks/sql/options.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,172 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"fmt" | ||||
| 	"time" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/logger" | ||||
| 	"go.unistack.org/micro/v3/meter" | ||||
| 	"go.unistack.org/micro/v3/tracer" | ||||
| ) | ||||
|  | ||||
| var ( | ||||
| 	// DefaultMeterStatsInterval holds default stats interval | ||||
| 	DefaultMeterStatsInterval = 5 * time.Second | ||||
| 	// DefaultLoggerObserver used to prepare labels for logger | ||||
| 	DefaultLoggerObserver = func(ctx context.Context, method string, query string, td time.Duration, err error) []interface{} { | ||||
| 		labels := []interface{}{"db.method", method, "took", fmt.Sprintf("%v", td)} | ||||
| 		if err != nil { | ||||
| 			labels = append(labels, "error", err.Error()) | ||||
| 		} | ||||
| 		if query != labelUnknown { | ||||
| 			labels = append(labels, "query", query) | ||||
| 		} | ||||
| 		return labels | ||||
| 	} | ||||
| ) | ||||
|  | ||||
| var ( | ||||
| 	MaxOpenConnections = "micro_sql_max_open_conn" | ||||
| 	OpenConnections    = "micro_sql_open_conn" | ||||
| 	InuseConnections   = "micro_sql_inuse_conn" | ||||
| 	IdleConnections    = "micro_sql_idle_conn" | ||||
| 	WaitConnections    = "micro_sql_waited_conn" | ||||
| 	BlockedSeconds     = "micro_sql_blocked_seconds" | ||||
| 	MaxIdleClosed      = "micro_sql_max_idle_closed" | ||||
| 	MaxIdletimeClosed  = "micro_sql_closed_max_idle" | ||||
| 	MaxLifetimeClosed  = "micro_sql_closed_max_lifetime" | ||||
|  | ||||
| 	meterRequestTotal               = "micro_sql_request_total" | ||||
| 	meterRequestLatencyMicroseconds = "micro_sql_latency_microseconds" | ||||
| 	meterRequestDurationSeconds     = "micro_sql_request_duration_seconds" | ||||
|  | ||||
| 	labelUnknown  = "unknown" | ||||
| 	labelQuery    = "db_statement" | ||||
| 	labelMethod   = "db_method" | ||||
| 	labelStatus   = "status" | ||||
| 	labelSuccess  = "success" | ||||
| 	labelFailure  = "failure" | ||||
| 	labelHost     = "db_host" | ||||
| 	labelDatabase = "db_name" | ||||
| ) | ||||
|  | ||||
| // Options struct holds wrapper options | ||||
| type Options struct { | ||||
| 	Logger             logger.Logger | ||||
| 	Meter              meter.Meter | ||||
| 	Tracer             tracer.Tracer | ||||
| 	DatabaseHost       string | ||||
| 	DatabaseName       string | ||||
| 	MeterStatsInterval time.Duration | ||||
| 	LoggerLevel        logger.Level | ||||
| 	LoggerEnabled      bool | ||||
| 	LoggerObserver     func(ctx context.Context, method string, name string, td time.Duration, err error) []interface{} | ||||
| } | ||||
|  | ||||
| // Option func signature | ||||
| type Option func(*Options) | ||||
|  | ||||
| // NewOptions create new Options struct from provided option slice | ||||
| func NewOptions(opts ...Option) Options { | ||||
| 	options := Options{ | ||||
| 		Logger:             logger.DefaultLogger, | ||||
| 		Meter:              meter.DefaultMeter, | ||||
| 		Tracer:             tracer.DefaultTracer, | ||||
| 		MeterStatsInterval: DefaultMeterStatsInterval, | ||||
| 		LoggerLevel:        logger.ErrorLevel, | ||||
| 		LoggerObserver:     DefaultLoggerObserver, | ||||
| 	} | ||||
| 	for _, o := range opts { | ||||
| 		o(&options) | ||||
| 	} | ||||
|  | ||||
| 	options.Meter = options.Meter.Clone( | ||||
| 		meter.Labels( | ||||
| 			labelHost, options.DatabaseHost, | ||||
| 			labelDatabase, options.DatabaseName, | ||||
| 		), | ||||
| 	) | ||||
|  | ||||
| 	options.Logger = options.Logger.Clone(logger.WithAddCallerSkipCount(1)) | ||||
|  | ||||
| 	return options | ||||
| } | ||||
|  | ||||
| // MetricInterval specifies stats interval for *sql.DB | ||||
| func MetricInterval(td time.Duration) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.MeterStatsInterval = td | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func DatabaseHost(host string) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.DatabaseHost = host | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func DatabaseName(name string) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.DatabaseName = name | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Meter passes meter.Meter to wrapper | ||||
| func Meter(m meter.Meter) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Meter = m | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Logger passes logger.Logger to wrapper | ||||
| func Logger(l logger.Logger) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Logger = l | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // LoggerEnabled enable sql logging | ||||
| func LoggerEnabled(b bool) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.LoggerEnabled = b | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // LoggerLevel passes logger.Level option | ||||
| func LoggerLevel(lvl logger.Level) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.LoggerLevel = lvl | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // LoggerObserver passes observer to fill logger fields | ||||
| func LoggerObserver(obs func(context.Context, string, string, time.Duration, error) []interface{}) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.LoggerObserver = obs | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Tracer passes tracer.Tracer to wrapper | ||||
| func Tracer(t tracer.Tracer) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Tracer = t | ||||
| 	} | ||||
| } | ||||
|  | ||||
| type queryNameKey struct{} | ||||
|  | ||||
| // QueryName passes query name to wrapper func | ||||
| func QueryName(ctx context.Context, name string) context.Context { | ||||
| 	if ctx == nil { | ||||
| 		ctx = context.Background() | ||||
| 	} | ||||
| 	return context.WithValue(ctx, queryNameKey{}, name) | ||||
| } | ||||
|  | ||||
| func getQueryName(ctx context.Context) string { | ||||
| 	if v, ok := ctx.Value(queryNameKey{}).(string); ok && v != labelUnknown { | ||||
| 		return v | ||||
| 	} | ||||
| 	return getCallerName() | ||||
| } | ||||
							
								
								
									
										41
									
								
								hooks/sql/stats.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										41
									
								
								hooks/sql/stats.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,41 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"database/sql" | ||||
| 	"time" | ||||
| ) | ||||
|  | ||||
| type Statser interface { | ||||
| 	Stats() sql.DBStats | ||||
| } | ||||
|  | ||||
| func NewStatsMeter(ctx context.Context, db Statser, opts ...Option) { | ||||
| 	options := NewOptions(opts...) | ||||
|  | ||||
| 	go func() { | ||||
| 		ticker := time.NewTicker(options.MeterStatsInterval) | ||||
| 		defer ticker.Stop() | ||||
|  | ||||
| 		for { | ||||
| 			select { | ||||
| 			case <-ctx.Done(): | ||||
| 				return | ||||
| 			case <-ticker.C: | ||||
| 				if db == nil { | ||||
| 					return | ||||
| 				} | ||||
| 				stats := db.Stats() | ||||
| 				options.Meter.Counter(MaxOpenConnections).Set(uint64(stats.MaxOpenConnections)) | ||||
| 				options.Meter.Counter(OpenConnections).Set(uint64(stats.OpenConnections)) | ||||
| 				options.Meter.Counter(InuseConnections).Set(uint64(stats.InUse)) | ||||
| 				options.Meter.Counter(IdleConnections).Set(uint64(stats.Idle)) | ||||
| 				options.Meter.Counter(WaitConnections).Set(uint64(stats.WaitCount)) | ||||
| 				options.Meter.FloatCounter(BlockedSeconds).Set(stats.WaitDuration.Seconds()) | ||||
| 				options.Meter.Counter(MaxIdleClosed).Set(uint64(stats.MaxIdleClosed)) | ||||
| 				options.Meter.Counter(MaxIdletimeClosed).Set(uint64(stats.MaxIdleTimeClosed)) | ||||
| 				options.Meter.Counter(MaxLifetimeClosed).Set(uint64(stats.MaxLifetimeClosed)) | ||||
| 			} | ||||
| 		} | ||||
| 	}() | ||||
| } | ||||
							
								
								
									
										287
									
								
								hooks/sql/stmt.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										287
									
								
								hooks/sql/stmt.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,287 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"database/sql/driver" | ||||
| 	"fmt" | ||||
| 	"time" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/hooks/requestid" | ||||
| 	"go.unistack.org/micro/v3/tracer" | ||||
| ) | ||||
|  | ||||
| var ( | ||||
| 	_ driver.Stmt              = (*wrapperStmt)(nil) | ||||
| 	_ driver.StmtQueryContext  = (*wrapperStmt)(nil) | ||||
| 	_ driver.StmtExecContext   = (*wrapperStmt)(nil) | ||||
| 	_ driver.NamedValueChecker = (*wrapperStmt)(nil) | ||||
| ) | ||||
|  | ||||
| // wrapperStmt defines a wrapper for driver.Stmt | ||||
| type wrapperStmt struct { | ||||
| 	stmt  driver.Stmt | ||||
| 	opts  Options | ||||
| 	query string | ||||
| 	ctx   context.Context | ||||
| } | ||||
|  | ||||
| // Close implements driver.Stmt Close | ||||
| func (w *wrapperStmt) Close() error { | ||||
| 	var ctx context.Context | ||||
| 	if w.ctx != nil { | ||||
| 		ctx = w.ctx | ||||
| 	} else { | ||||
| 		ctx = context.Background() | ||||
| 	} | ||||
| 	_ = ctx | ||||
| 	labels := []string{labelMethod, "Close"} | ||||
| 	ts := time.Now() | ||||
| 	err := w.stmt.Close() | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 	} else { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	} | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "Close", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| // NumInput implements driver.Stmt NumInput | ||||
| func (w *wrapperStmt) NumInput() int { | ||||
| 	return w.stmt.NumInput() | ||||
| } | ||||
|  | ||||
| // CheckNamedValue implements driver.NamedValueChecker | ||||
| func (w *wrapperStmt) CheckNamedValue(v *driver.NamedValue) error { | ||||
| 	s, ok := w.stmt.(driver.NamedValueChecker) | ||||
| 	if !ok { | ||||
| 		return driver.ErrSkip | ||||
| 	} | ||||
| 	return s.CheckNamedValue(v) | ||||
| } | ||||
|  | ||||
| // Exec implements driver.Stmt Exec | ||||
| func (w *wrapperStmt) Exec(args []driver.Value) (driver.Result, error) { | ||||
| 	var ctx context.Context | ||||
| 	if w.ctx != nil { | ||||
| 		ctx = w.ctx | ||||
| 	} else { | ||||
| 		ctx = context.Background() | ||||
| 	} | ||||
| 	_ = ctx | ||||
| 	labels := []string{labelMethod, "Exec"} | ||||
| 	ts := time.Now() | ||||
| 	res, err := w.stmt.Exec(args) // nolint:staticcheck | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 	} else { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	} | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "Exec", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return res, err | ||||
| } | ||||
|  | ||||
| // Query implements driver.Stmt Query | ||||
| func (w *wrapperStmt) Query(args []driver.Value) (driver.Rows, error) { | ||||
| 	var ctx context.Context | ||||
| 	if w.ctx != nil { | ||||
| 		ctx = w.ctx | ||||
| 	} else { | ||||
| 		ctx = context.Background() | ||||
| 	} | ||||
| 	_ = ctx | ||||
| 	labels := []string{labelMethod, "Query"} | ||||
| 	ts := time.Now() | ||||
| 	rows, err := w.stmt.Query(args) // nolint:staticcheck | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 	} else { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	} | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "Query", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return rows, err | ||||
| } | ||||
|  | ||||
| // ColumnConverter implements driver.ColumnConverter | ||||
| func (w *wrapperStmt) ColumnConverter(idx int) driver.ValueConverter { | ||||
| 	s, ok := w.stmt.(driver.ColumnConverter) // nolint:staticcheck | ||||
| 	if !ok { | ||||
| 		return nil | ||||
| 	} | ||||
| 	return s.ColumnConverter(idx) | ||||
| } | ||||
|  | ||||
| // ExecContext implements driver.StmtExecContext ExecContext | ||||
| func (w *wrapperStmt) ExecContext(ctx context.Context, args []driver.NamedValue) (driver.Result, error) { | ||||
| 	var nctx context.Context | ||||
| 	var span tracer.Span | ||||
|  | ||||
| 	name := getQueryName(ctx) | ||||
| 	if w.ctx != nil { | ||||
| 		nctx, span = w.opts.Tracer.Start(w.ctx, "sdk.database", tracer.WithSpanKind(tracer.SpanKindClient)) | ||||
| 	} else { | ||||
| 		nctx, span = w.opts.Tracer.Start(ctx, "sdk.database", tracer.WithSpanKind(tracer.SpanKindClient)) | ||||
| 	} | ||||
| 	span.AddLabels("db.method", "ExecContext") | ||||
| 	span.AddLabels("db.statement", name) | ||||
| 	defer span.Finish() | ||||
| 	if len(args) > 0 { | ||||
| 		span.AddLabels("db.args", fmt.Sprintf("%v", namedValueToLabels(args))) | ||||
| 	} | ||||
| 	if id, ok := ctx.Value(requestid.XRequestIDKey{}).(string); ok { | ||||
| 		span.AddLabels("x-request-id", id) | ||||
| 	} | ||||
| 	labels := []string{labelMethod, "ExecContext", labelQuery, name} | ||||
|  | ||||
| 	if conn, ok := w.stmt.(driver.StmtExecContext); ok { | ||||
| 		ts := time.Now() | ||||
| 		res, err := conn.ExecContext(nctx, args) | ||||
| 		td := time.Since(ts) | ||||
| 		te := td.Seconds() | ||||
| 		if err != nil { | ||||
| 			w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 			span.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 		} else { | ||||
| 			w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 		} | ||||
| 		w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 		w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 		/* | ||||
| 			if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 				w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "ExecContext", name, td, err)...) | ||||
| 			} | ||||
| 		*/ | ||||
| 		return res, err | ||||
| 	} | ||||
|  | ||||
| 	values, err := namedValueToValue(args) | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 		span.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 		/* | ||||
| 			if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 				w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "ExecContext", name, 0, err)...) | ||||
| 			} | ||||
| 		*/ | ||||
| 		return nil, err | ||||
| 	} | ||||
| 	ts := time.Now() | ||||
| 	res, err := w.Exec(values) // nolint:staticcheck | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 		span.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 	} else { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	} | ||||
|  | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "ExecContext", name, td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return res, err | ||||
| } | ||||
|  | ||||
| // QueryContext implements driver.StmtQueryContext StmtQueryContext | ||||
| func (w *wrapperStmt) QueryContext(ctx context.Context, args []driver.NamedValue) (driver.Rows, error) { | ||||
| 	var nctx context.Context | ||||
| 	var span tracer.Span | ||||
|  | ||||
| 	name := getQueryName(ctx) | ||||
| 	if w.ctx != nil { | ||||
| 		nctx, span = w.opts.Tracer.Start(w.ctx, "sdk.database", tracer.WithSpanKind(tracer.SpanKindClient)) | ||||
| 	} else { | ||||
| 		nctx, span = w.opts.Tracer.Start(ctx, "sdk.database", tracer.WithSpanKind(tracer.SpanKindClient)) | ||||
| 	} | ||||
| 	span.AddLabels("db.method", "QueryContext") | ||||
| 	span.AddLabels("db.statement", name) | ||||
| 	defer span.Finish() | ||||
| 	if len(args) > 0 { | ||||
| 		span.AddLabels("db.args", fmt.Sprintf("%v", namedValueToLabels(args))) | ||||
| 	} | ||||
| 	if id, ok := ctx.Value(requestid.XRequestIDKey{}).(string); ok { | ||||
| 		span.AddLabels("x-request-id", id) | ||||
| 	} | ||||
| 	labels := []string{labelMethod, "QueryContext", labelQuery, name} | ||||
| 	if conn, ok := w.stmt.(driver.StmtQueryContext); ok { | ||||
| 		ts := time.Now() | ||||
| 		rows, err := conn.QueryContext(nctx, args) | ||||
| 		td := time.Since(ts) | ||||
| 		te := td.Seconds() | ||||
| 		if err != nil { | ||||
| 			w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 			span.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 		} else { | ||||
| 			w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 		} | ||||
|  | ||||
| 		w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 		w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 		/* | ||||
| 			if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 				w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "QueryContext", name, td, err)...) | ||||
| 			} | ||||
| 		*/ | ||||
| 		return rows, err | ||||
| 	} | ||||
|  | ||||
| 	values, err := namedValueToValue(args) | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
|  | ||||
| 		span.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 		/* | ||||
| 			if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 				w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "QueryContext", name, 0, err)...) | ||||
| 			} | ||||
| 		*/ | ||||
| 		return nil, err | ||||
| 	} | ||||
| 	ts := time.Now() | ||||
| 	rows, err := w.Query(values) // nolint:staticcheck | ||||
| 	td := time.Since(ts) | ||||
| 	te := td.Seconds() | ||||
| 	if err != nil { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelFailure)...).Inc() | ||||
| 		span.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 	} else { | ||||
| 		w.opts.Meter.Counter(meterRequestTotal, append(labels, labelStatus, labelSuccess)...).Inc() | ||||
| 	} | ||||
|  | ||||
| 	w.opts.Meter.Summary(meterRequestLatencyMicroseconds, labels...).Update(te) | ||||
| 	w.opts.Meter.Histogram(meterRequestDurationSeconds, labels...).Update(te) | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(ctx, "QueryContext", name, td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	return rows, err | ||||
| } | ||||
							
								
								
									
										63
									
								
								hooks/sql/tx.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										63
									
								
								hooks/sql/tx.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,63 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"database/sql/driver" | ||||
| 	"time" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/tracer" | ||||
| ) | ||||
|  | ||||
| var _ driver.Tx = (*wrapperTx)(nil) | ||||
|  | ||||
| // wrapperTx defines a wrapper for driver.Tx | ||||
| type wrapperTx struct { | ||||
| 	tx   driver.Tx | ||||
| 	span tracer.Span | ||||
| 	opts Options | ||||
| 	ctx  context.Context | ||||
| } | ||||
|  | ||||
| // Commit implements driver.Tx Commit | ||||
| func (w *wrapperTx) Commit() error { | ||||
| 	ts := time.Now() | ||||
| 	err := w.tx.Commit() | ||||
| 	td := time.Since(ts) | ||||
| 	_ = td | ||||
| 	if w.span != nil { | ||||
| 		if err != nil { | ||||
| 			w.span.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 		} | ||||
| 		w.span.Finish() | ||||
| 	} | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(w.ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(w.ctx, "Commit", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	w.ctx = nil | ||||
|  | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| // Rollback implements driver.Tx Rollback | ||||
| func (w *wrapperTx) Rollback() error { | ||||
| 	ts := time.Now() | ||||
| 	err := w.tx.Rollback() | ||||
| 	td := time.Since(ts) | ||||
| 	_ = td | ||||
| 	if w.span != nil { | ||||
| 		if err != nil { | ||||
| 			w.span.SetStatus(tracer.SpanStatusError, err.Error()) | ||||
| 		} | ||||
| 		w.span.Finish() | ||||
| 	} | ||||
| 	/* | ||||
| 		if w.opts.LoggerEnabled && w.opts.Logger.V(w.opts.LoggerLevel) { | ||||
| 			w.opts.Logger.Log(w.ctx, w.opts.LoggerLevel, w.opts.LoggerObserver(w.ctx, "Rollback", getCallerName(), td, err)...) | ||||
| 		} | ||||
| 	*/ | ||||
| 	w.ctx = nil | ||||
|  | ||||
| 	return err | ||||
| } | ||||
							
								
								
									
										19
									
								
								hooks/sql/wrap.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										19
									
								
								hooks/sql/wrap.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,19 @@ | ||||
| package sql | ||||
|  | ||||
| import ( | ||||
| 	"database/sql/driver" | ||||
| ) | ||||
|  | ||||
| /* | ||||
| func wrapDriver(d driver.Driver, opts Options) driver.Driver { | ||||
| 	if _, ok := d.(driver.DriverContext); ok { | ||||
| 		return &wrapperDriver{driver: d, opts: opts} | ||||
| 	} | ||||
| 	return struct{ driver.Driver }{&wrapperDriver{driver: d, opts: opts}} | ||||
| } | ||||
| */ | ||||
|  | ||||
| // WrapConn allows an existing driver.Conn to be wrapped. | ||||
| func WrapConn(c driver.Conn, opts ...Option) driver.Conn { | ||||
| 	return wrapConn(c, NewOptions(opts...)) | ||||
| } | ||||
							
								
								
									
										20699
									
								
								hooks/sql/wrap_gen.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										20699
									
								
								hooks/sql/wrap_gen.go
									
									
									
									
									
										Normal file
									
								
							
										
											
												File diff suppressed because it is too large
												Load Diff
											
										
									
								
							
							
								
								
									
										194
									
								
								hooks/validator/validator.go
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										194
									
								
								hooks/validator/validator.go
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,194 @@ | ||||
| package validator | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/client" | ||||
| 	"go.unistack.org/micro/v3/errors" | ||||
| 	"go.unistack.org/micro/v3/server" | ||||
| ) | ||||
|  | ||||
| var ( | ||||
| 	DefaultClientErrorFunc = func(req client.Request, rsp interface{}, err error) error { | ||||
| 		if rsp != nil { | ||||
| 			return errors.BadGateway(req.Service(), "%v", err) | ||||
| 		} | ||||
| 		return errors.BadRequest(req.Service(), "%v", err) | ||||
| 	} | ||||
|  | ||||
| 	DefaultServerErrorFunc = func(req server.Request, rsp interface{}, err error) error { | ||||
| 		if rsp != nil { | ||||
| 			return errors.BadGateway(req.Service(), "%v", err) | ||||
| 		} | ||||
| 		return errors.BadRequest(req.Service(), "%v", err) | ||||
| 	} | ||||
|  | ||||
| 	DefaultPublishErrorFunc = func(msg client.Message, err error) error { | ||||
| 		return errors.BadRequest(msg.Topic(), "%v", err) | ||||
| 	} | ||||
|  | ||||
| 	DefaultSubscribeErrorFunc = func(msg server.Message, err error) error { | ||||
| 		return errors.BadRequest(msg.Topic(), "%v", err) | ||||
| 	} | ||||
| ) | ||||
|  | ||||
| type ( | ||||
| 	ClientErrorFunc    func(client.Request, interface{}, error) error | ||||
| 	ServerErrorFunc    func(server.Request, interface{}, error) error | ||||
| 	PublishErrorFunc   func(client.Message, error) error | ||||
| 	SubscribeErrorFunc func(server.Message, error) error | ||||
| ) | ||||
|  | ||||
| // Options struct holds wrapper options | ||||
| type Options struct { | ||||
| 	ClientErrorFn          ClientErrorFunc | ||||
| 	ServerErrorFn          ServerErrorFunc | ||||
| 	PublishErrorFn         PublishErrorFunc | ||||
| 	SubscribeErrorFn       SubscribeErrorFunc | ||||
| 	ClientValidateResponse bool | ||||
| 	ServerValidateResponse bool | ||||
| } | ||||
|  | ||||
| // Option func signature | ||||
| type Option func(*Options) | ||||
|  | ||||
| func ClientValidateResponse(b bool) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.ClientValidateResponse = b | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func ServerValidateResponse(b bool) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.ClientValidateResponse = b | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func ClientReqErrorFn(fn ClientErrorFunc) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.ClientErrorFn = fn | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func ServerErrorFn(fn ServerErrorFunc) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.ServerErrorFn = fn | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func PublishErrorFn(fn PublishErrorFunc) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.PublishErrorFn = fn | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func SubscribeErrorFn(fn SubscribeErrorFunc) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.SubscribeErrorFn = fn | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func NewOptions(opts ...Option) Options { | ||||
| 	options := Options{ | ||||
| 		ClientErrorFn:    DefaultClientErrorFunc, | ||||
| 		ServerErrorFn:    DefaultServerErrorFunc, | ||||
| 		PublishErrorFn:   DefaultPublishErrorFunc, | ||||
| 		SubscribeErrorFn: DefaultSubscribeErrorFunc, | ||||
| 	} | ||||
| 	for _, o := range opts { | ||||
| 		o(&options) | ||||
| 	} | ||||
| 	return options | ||||
| } | ||||
|  | ||||
| func NewHook(opts ...Option) *hook { | ||||
| 	return &hook{opts: NewOptions(opts...)} | ||||
| } | ||||
|  | ||||
| type validator interface { | ||||
| 	Validate() error | ||||
| } | ||||
|  | ||||
| type hook struct { | ||||
| 	opts Options | ||||
| } | ||||
|  | ||||
| func (w *hook) ClientCall(next client.FuncCall) client.FuncCall { | ||||
| 	return func(ctx context.Context, req client.Request, rsp interface{}, opts ...client.CallOption) error { | ||||
| 		if v, ok := req.Body().(validator); ok { | ||||
| 			if err := v.Validate(); err != nil { | ||||
| 				return w.opts.ClientErrorFn(req, nil, err) | ||||
| 			} | ||||
| 		} | ||||
| 		err := next(ctx, req, rsp, opts...) | ||||
| 		if v, ok := rsp.(validator); ok && w.opts.ClientValidateResponse { | ||||
| 			if verr := v.Validate(); verr != nil { | ||||
| 				return w.opts.ClientErrorFn(req, rsp, verr) | ||||
| 			} | ||||
| 		} | ||||
| 		return err | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ClientStream(next client.FuncStream) client.FuncStream { | ||||
| 	return func(ctx context.Context, req client.Request, opts ...client.CallOption) (client.Stream, error) { | ||||
| 		if v, ok := req.Body().(validator); ok { | ||||
| 			if err := v.Validate(); err != nil { | ||||
| 				return nil, w.opts.ClientErrorFn(req, nil, err) | ||||
| 			} | ||||
| 		} | ||||
| 		return next(ctx, req, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ClientPublish(next client.FuncPublish) client.FuncPublish { | ||||
| 	return func(ctx context.Context, msg client.Message, opts ...client.PublishOption) error { | ||||
| 		if v, ok := msg.Payload().(validator); ok { | ||||
| 			if err := v.Validate(); err != nil { | ||||
| 				return w.opts.PublishErrorFn(msg, err) | ||||
| 			} | ||||
| 		} | ||||
| 		return next(ctx, msg, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ClientBatchPublish(next client.FuncBatchPublish) client.FuncBatchPublish { | ||||
| 	return func(ctx context.Context, msgs []client.Message, opts ...client.PublishOption) error { | ||||
| 		for _, msg := range msgs { | ||||
| 			if v, ok := msg.Payload().(validator); ok { | ||||
| 				if err := v.Validate(); err != nil { | ||||
| 					return w.opts.PublishErrorFn(msg, err) | ||||
| 				} | ||||
| 			} | ||||
| 		} | ||||
| 		return next(ctx, msgs, opts...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ServerHandler(next server.FuncHandler) server.FuncHandler { | ||||
| 	return func(ctx context.Context, req server.Request, rsp interface{}) error { | ||||
| 		if v, ok := req.Body().(validator); ok { | ||||
| 			if err := v.Validate(); err != nil { | ||||
| 				return w.opts.ServerErrorFn(req, nil, err) | ||||
| 			} | ||||
| 		} | ||||
| 		err := next(ctx, req, rsp) | ||||
| 		if v, ok := rsp.(validator); ok && w.opts.ServerValidateResponse { | ||||
| 			if verr := v.Validate(); verr != nil { | ||||
| 				return w.opts.ServerErrorFn(req, rsp, verr) | ||||
| 			} | ||||
| 		} | ||||
| 		return err | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *hook) ServerSubscriber(next server.FuncSubHandler) server.FuncSubHandler { | ||||
| 	return func(ctx context.Context, msg server.Message) error { | ||||
| 		if v, ok := msg.Body().(validator); ok { | ||||
| 			if err := v.Validate(); err != nil { | ||||
| 				return w.opts.SubscribeErrorFn(msg, err) | ||||
| 			} | ||||
| 		} | ||||
| 		return next(ctx, msg) | ||||
| 	} | ||||
| } | ||||
| @@ -13,6 +13,15 @@ func FromContext(ctx context.Context) (Logger, bool) { | ||||
| 	return l, ok | ||||
| } | ||||
|  | ||||
| // MustContext returns logger from passed context or DefaultLogger if empty | ||||
| func MustContext(ctx context.Context) Logger { | ||||
| 	l, ok := FromContext(ctx) | ||||
| 	if !ok { | ||||
| 		panic("missing logger") | ||||
| 	} | ||||
| 	return l | ||||
| } | ||||
|  | ||||
| // NewContext stores logger into passed context | ||||
| func NewContext(ctx context.Context, l Logger) context.Context { | ||||
| 	if ctx == nil { | ||||
|   | ||||
| @@ -4,18 +4,20 @@ package logger | ||||
| type Level int8 | ||||
|  | ||||
| const ( | ||||
| 	// TraceLevel level usually used to find bugs, very verbose | ||||
| 	// TraceLevel usually used to find bugs, very verbose | ||||
| 	TraceLevel Level = iota - 2 | ||||
| 	// DebugLevel level used only when enabled debugging | ||||
| 	// DebugLevel used only when enabled debugging | ||||
| 	DebugLevel | ||||
| 	// InfoLevel level used for general info about what's going on inside the application | ||||
| 	// InfoLevel used for general info about what's going on inside the application | ||||
| 	InfoLevel | ||||
| 	// WarnLevel level used for non-critical entries | ||||
| 	// WarnLevel used for non-critical entries | ||||
| 	WarnLevel | ||||
| 	// ErrorLevel level used for errors that should definitely be noted | ||||
| 	// ErrorLevel used for errors that should definitely be noted | ||||
| 	ErrorLevel | ||||
| 	// FatalLevel level used for critical errors and then calls `os.Exit(1)` | ||||
| 	// FatalLevel used for critical errors and then calls `os.Exit(1)` | ||||
| 	FatalLevel | ||||
| 	// NoneLevel used to disable logging | ||||
| 	NoneLevel | ||||
| ) | ||||
|  | ||||
| // String returns logger level string representation | ||||
| @@ -33,6 +35,8 @@ func (l Level) String() string { | ||||
| 		return "error" | ||||
| 	case FatalLevel: | ||||
| 		return "fatal" | ||||
| 	case NoneLevel: | ||||
| 		return "none" | ||||
| 	} | ||||
| 	return "info" | ||||
| } | ||||
| @@ -58,6 +62,8 @@ func ParseLevel(lvl string) Level { | ||||
| 		return ErrorLevel | ||||
| 	case FatalLevel.String(): | ||||
| 		return FatalLevel | ||||
| 	case NoneLevel.String(): | ||||
| 		return NoneLevel | ||||
| 	} | ||||
| 	return InfoLevel | ||||
| } | ||||
|   | ||||
							
								
								
									
										139
									
								
								logger/logger.go
									
									
									
									
									
								
							
							
						
						
									
										139
									
								
								logger/logger.go
									
									
									
									
									
								
							| @@ -1,5 +1,5 @@ | ||||
| // Package logger provides a log interface | ||||
| package logger // import "go.unistack.org/micro/v3/logger" | ||||
| package logger | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| @@ -11,11 +11,9 @@ var DefaultContextAttrFuncs []ContextAttrFunc | ||||
|  | ||||
| var ( | ||||
| 	// DefaultLogger variable | ||||
| 	DefaultLogger Logger = NewLogger() | ||||
| 	DefaultLogger = NewLogger() | ||||
| 	// DefaultLevel used by logger | ||||
| 	DefaultLevel = InfoLevel | ||||
| 	// DefaultCallerSkipCount used by logger | ||||
| 	DefaultCallerSkipCount = 2 | ||||
| ) | ||||
|  | ||||
| // Logger is a generic logging interface | ||||
| @@ -33,33 +31,19 @@ type Logger interface { | ||||
| 	// Fields set fields to always be logged with keyval pairs | ||||
| 	Fields(fields ...interface{}) Logger | ||||
| 	// Info level message | ||||
| 	Info(ctx context.Context, args ...interface{}) | ||||
| 	Info(ctx context.Context, msg string, args ...interface{}) | ||||
| 	// Trace level message | ||||
| 	Trace(ctx context.Context, args ...interface{}) | ||||
| 	Trace(ctx context.Context, msg string, args ...interface{}) | ||||
| 	// Debug level message | ||||
| 	Debug(ctx context.Context, args ...interface{}) | ||||
| 	Debug(ctx context.Context, msg string, args ...interface{}) | ||||
| 	// Warn level message | ||||
| 	Warn(ctx context.Context, args ...interface{}) | ||||
| 	Warn(ctx context.Context, msg string, args ...interface{}) | ||||
| 	// Error level message | ||||
| 	Error(ctx context.Context, args ...interface{}) | ||||
| 	Error(ctx context.Context, msg string, args ...interface{}) | ||||
| 	// Fatal level message | ||||
| 	Fatal(ctx context.Context, args ...interface{}) | ||||
| 	// Infof level message | ||||
| 	Infof(ctx context.Context, msg string, args ...interface{}) | ||||
| 	// Tracef level message | ||||
| 	Tracef(ctx context.Context, msg string, args ...interface{}) | ||||
| 	// Debug level message | ||||
| 	Debugf(ctx context.Context, msg string, args ...interface{}) | ||||
| 	// Warn level message | ||||
| 	Warnf(ctx context.Context, msg string, args ...interface{}) | ||||
| 	// Error level message | ||||
| 	Errorf(ctx context.Context, msg string, args ...interface{}) | ||||
| 	// Fatal level message | ||||
| 	Fatalf(ctx context.Context, msg string, args ...interface{}) | ||||
| 	Fatal(ctx context.Context, msg string, args ...interface{}) | ||||
| 	// Log logs message with needed level | ||||
| 	Log(ctx context.Context, level Level, args ...interface{}) | ||||
| 	// Logf logs message with needed level | ||||
| 	Logf(ctx context.Context, level Level, msg string, args ...interface{}) | ||||
| 	Log(ctx context.Context, level Level, msg string, args ...interface{}) | ||||
| 	// Name returns broker instance name | ||||
| 	Name() string | ||||
| 	// String returns the type of logger | ||||
| @@ -68,108 +52,3 @@ type Logger interface { | ||||
|  | ||||
| // Field contains keyval pair | ||||
| type Field interface{} | ||||
|  | ||||
| // Info writes msg to default logger on info level | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Info(ctx context.Context, args ...interface{}) { | ||||
| 	DefaultLogger.Clone(WithCallerSkipCount(DefaultCallerSkipCount+1)).Info(ctx, args...) | ||||
| } | ||||
|  | ||||
| // Error writes msg to default logger on error level | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Error(ctx context.Context, args ...interface{}) { | ||||
| 	DefaultLogger.Clone(WithCallerSkipCount(DefaultCallerSkipCount+1)).Error(ctx, args...) | ||||
| } | ||||
|  | ||||
| // Debug writes msg to default logger on debug level | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Debug(ctx context.Context, args ...interface{}) { | ||||
| 	DefaultLogger.Clone(WithCallerSkipCount(DefaultCallerSkipCount+1)).Debug(ctx, args...) | ||||
| } | ||||
|  | ||||
| // Warn writes msg to default logger on warn level | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Warn(ctx context.Context, args ...interface{}) { | ||||
| 	DefaultLogger.Clone(WithCallerSkipCount(DefaultCallerSkipCount+1)).Warn(ctx, args...) | ||||
| } | ||||
|  | ||||
| // Trace writes msg to default logger on trace level | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Trace(ctx context.Context, args ...interface{}) { | ||||
| 	DefaultLogger.Clone(WithCallerSkipCount(DefaultCallerSkipCount+1)).Trace(ctx, args...) | ||||
| } | ||||
|  | ||||
| // Fatal writes msg to default logger on fatal level | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Fatal(ctx context.Context, args ...interface{}) { | ||||
| 	DefaultLogger.Clone(WithCallerSkipCount(DefaultCallerSkipCount+1)).Fatal(ctx, args...) | ||||
| } | ||||
|  | ||||
| // Infof writes formatted msg to default logger on info level | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Infof(ctx context.Context, msg string, args ...interface{}) { | ||||
| 	DefaultLogger.Clone(WithCallerSkipCount(DefaultCallerSkipCount+1)).Infof(ctx, msg, args...) | ||||
| } | ||||
|  | ||||
| // Errorf writes formatted msg to default logger on error level | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Errorf(ctx context.Context, msg string, args ...interface{}) { | ||||
| 	DefaultLogger.Clone(WithCallerSkipCount(DefaultCallerSkipCount+1)).Errorf(ctx, msg, args...) | ||||
| } | ||||
|  | ||||
| // Debugf writes formatted msg to default logger on debug level | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Debugf(ctx context.Context, msg string, args ...interface{}) { | ||||
| 	DefaultLogger.Clone(WithCallerSkipCount(DefaultCallerSkipCount+1)).Debugf(ctx, msg, args...) | ||||
| } | ||||
|  | ||||
| // Warnf writes formatted msg to default logger on warn level | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Warnf(ctx context.Context, msg string, args ...interface{}) { | ||||
| 	DefaultLogger.Clone(WithCallerSkipCount(DefaultCallerSkipCount+1)).Warnf(ctx, msg, args...) | ||||
| } | ||||
|  | ||||
| // Tracef writes formatted msg to default logger on trace level | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Tracef(ctx context.Context, msg string, args ...interface{}) { | ||||
| 	DefaultLogger.Clone(WithCallerSkipCount(DefaultCallerSkipCount+1)).Tracef(ctx, msg, args...) | ||||
| } | ||||
|  | ||||
| // Fatalf writes formatted msg to default logger on fatal level | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Fatalf(ctx context.Context, msg string, args ...interface{}) { | ||||
| 	DefaultLogger.Clone(WithCallerSkipCount(DefaultCallerSkipCount+1)).Fatalf(ctx, msg, args...) | ||||
| } | ||||
|  | ||||
| // V returns true if passed level enabled in default logger | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func V(level Level) bool { | ||||
| 	return DefaultLogger.V(level) | ||||
| } | ||||
|  | ||||
| // Init initialize logger | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Init(opts ...Option) error { | ||||
| 	return DefaultLogger.Init(opts...) | ||||
| } | ||||
|  | ||||
| // Fields create logger with specific fields | ||||
| // | ||||
| // Deprecated: Dont use logger methods directly, use instance of logger to avoid additional allocations | ||||
| func Fields(fields ...interface{}) Logger { | ||||
| 	return DefaultLogger.Fields(fields...) | ||||
| } | ||||
|   | ||||
| @@ -4,12 +4,17 @@ import ( | ||||
| 	"context" | ||||
| ) | ||||
|  | ||||
| const ( | ||||
| 	defaultCallerSkipCount = 2 | ||||
| ) | ||||
|  | ||||
| type noopLogger struct { | ||||
| 	opts Options | ||||
| } | ||||
|  | ||||
| func NewLogger(opts ...Option) Logger { | ||||
| 	options := NewOptions(opts...) | ||||
| 	options.CallerSkipCount = defaultCallerSkipCount | ||||
| 	return &noopLogger{opts: options} | ||||
| } | ||||
|  | ||||
| @@ -51,44 +56,23 @@ func (l *noopLogger) String() string { | ||||
| 	return "noop" | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Log(ctx context.Context, lvl Level, attrs ...interface{}) { | ||||
| func (l *noopLogger) Log(ctx context.Context, lvl Level, msg string, attrs ...interface{}) { | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Info(ctx context.Context, attrs ...interface{}) { | ||||
| func (l *noopLogger) Info(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Debug(ctx context.Context, attrs ...interface{}) { | ||||
| func (l *noopLogger) Debug(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Error(ctx context.Context, attrs ...interface{}) { | ||||
| func (l *noopLogger) Error(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Trace(ctx context.Context, attrs ...interface{}) { | ||||
| func (l *noopLogger) Trace(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Warn(ctx context.Context, attrs ...interface{}) { | ||||
| func (l *noopLogger) Warn(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Fatal(ctx context.Context, attrs ...interface{}) { | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Logf(ctx context.Context, lvl Level, msg string, attrs ...interface{}) { | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Infof(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Debugf(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Errorf(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Tracef(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Warnf(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| } | ||||
|  | ||||
| func (l *noopLogger) Fatalf(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| func (l *noopLogger) Fatal(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| } | ||||
|   | ||||
| @@ -5,7 +5,10 @@ import ( | ||||
| 	"io" | ||||
| 	"log/slog" | ||||
| 	"os" | ||||
| 	"slices" | ||||
| 	"time" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/meter" | ||||
| ) | ||||
|  | ||||
| // Option func signature | ||||
| @@ -13,18 +16,6 @@ type Option func(*Options) | ||||
|  | ||||
| // Options holds logger options | ||||
| type Options struct { | ||||
| 	// Out holds the output writer | ||||
| 	Out io.Writer | ||||
| 	// Context holds exernal options | ||||
| 	Context context.Context | ||||
| 	// Name holds the logger name | ||||
| 	Name string | ||||
| 	// Fields holds additional metadata | ||||
| 	Fields []interface{} | ||||
| 	// CallerSkipCount number of frmaes to skip | ||||
| 	CallerSkipCount int | ||||
| 	// ContextAttrFuncs contains funcs that executed before log func on context | ||||
| 	ContextAttrFuncs []ContextAttrFunc | ||||
| 	// TimeKey is the key used for the time of the log call | ||||
| 	TimeKey string | ||||
| 	// LevelKey is the key used for the level of the log call | ||||
| @@ -37,14 +28,36 @@ type Options struct { | ||||
| 	SourceKey string | ||||
| 	// StacktraceKey is the key used for the stacktrace | ||||
| 	StacktraceKey string | ||||
| 	// AddStacktrace controls writing of stacktaces on error | ||||
| 	AddStacktrace bool | ||||
| 	// AddSource enabled writing source file and position in log | ||||
| 	AddSource bool | ||||
| 	// The logging level the logger should log | ||||
| 	Level Level | ||||
| 	// Name holds the logger name | ||||
| 	Name string | ||||
| 	// Out holds the output writer | ||||
| 	Out io.Writer | ||||
| 	// Context holds exernal options | ||||
| 	Context context.Context | ||||
| 	// Meter used to count logs for specific level | ||||
| 	Meter meter.Meter | ||||
| 	// TimeFunc used to obtain current time | ||||
| 	TimeFunc func() time.Time | ||||
| 	// Fields holds additional metadata | ||||
| 	Fields []interface{} | ||||
| 	// ContextAttrFuncs contains funcs that executed before log func on context | ||||
| 	ContextAttrFuncs []ContextAttrFunc | ||||
| 	// callerSkipCount number of frmaes to skip | ||||
| 	CallerSkipCount int | ||||
| 	// The logging level the logger should log | ||||
| 	Level Level | ||||
| 	// AddSource enabled writing source file and position in log | ||||
| 	AddSource bool | ||||
| 	// AddStacktrace controls writing of stacktaces on error | ||||
| 	AddStacktrace bool | ||||
| 	// DedupKeys deduplicate keys in log output | ||||
| 	DedupKeys bool | ||||
| 	// FatalFinalizers runs in order in [logger.Fatal] method | ||||
| 	FatalFinalizers []func(context.Context) | ||||
| } | ||||
|  | ||||
| var DefaultFatalFinalizer = func(ctx context.Context) { | ||||
| 	os.Exit(1) | ||||
| } | ||||
|  | ||||
| // NewOptions creates new options struct | ||||
| @@ -53,11 +66,12 @@ func NewOptions(opts ...Option) Options { | ||||
| 		Level:            DefaultLevel, | ||||
| 		Fields:           make([]interface{}, 0, 6), | ||||
| 		Out:              os.Stderr, | ||||
| 		CallerSkipCount:  DefaultCallerSkipCount, | ||||
| 		Context:          context.Background(), | ||||
| 		ContextAttrFuncs: DefaultContextAttrFuncs, | ||||
| 		AddSource:        true, | ||||
| 		TimeFunc:         time.Now, | ||||
| 		Meter:            meter.DefaultMeter, | ||||
| 		FatalFinalizers:  []func(context.Context){DefaultFatalFinalizer}, | ||||
| 	} | ||||
|  | ||||
| 	WithMicroKeys()(&options) | ||||
| @@ -69,13 +83,50 @@ func NewOptions(opts ...Option) Options { | ||||
| 	return options | ||||
| } | ||||
|  | ||||
| // WithContextAttrFuncs appends default funcs for the context arrts filler | ||||
| // WithFatalFinalizers set logger.Fatal finalizers | ||||
| func WithFatalFinalizers(fncs ...func(context.Context)) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.FatalFinalizers = fncs | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithContextAttrFuncs appends default funcs for the context attrs filler | ||||
| func WithContextAttrFuncs(fncs ...ContextAttrFunc) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.ContextAttrFuncs = append(o.ContextAttrFuncs, fncs...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithDedupKeys dont log duplicate keys | ||||
| func WithDedupKeys(b bool) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.DedupKeys = b | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithAddFields add fields for the logger | ||||
| func WithAddFields(fields ...interface{}) Option { | ||||
| 	return func(o *Options) { | ||||
| 		if o.DedupKeys { | ||||
| 			for i := 0; i < len(o.Fields); i += 2 { | ||||
| 				for j := 0; j < len(fields); j += 2 { | ||||
| 					iv, iok := o.Fields[i].(string) | ||||
| 					jv, jok := fields[j].(string) | ||||
| 					if iok && jok && iv == jv { | ||||
| 						o.Fields[i+1] = fields[j+1] | ||||
| 						fields = slices.Delete(fields, j, j+2) | ||||
| 					} | ||||
| 				} | ||||
| 			} | ||||
| 			if len(fields) > 0 { | ||||
| 				o.Fields = append(o.Fields, fields...) | ||||
| 			} | ||||
| 		} else { | ||||
| 			o.Fields = append(o.Fields, fields...) | ||||
| 		} | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithFields set default fields for the logger | ||||
| func WithFields(fields ...interface{}) Option { | ||||
| 	return func(o *Options) { | ||||
| @@ -97,27 +148,20 @@ func WithOutput(out io.Writer) Option { | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WitAddStacktrace controls writing stacktrace on error | ||||
| // WithAddStacktrace controls writing stacktrace on error | ||||
| func WithAddStacktrace(v bool) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.AddStacktrace = v | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WitAddSource controls writing source file and pos in log | ||||
| // WithAddSource controls writing source file and pos in log | ||||
| func WithAddSource(v bool) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.AddSource = v | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithCallerSkipCount set frame count to skip | ||||
| func WithCallerSkipCount(c int) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.CallerSkipCount = c | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithContext set context | ||||
| func WithContext(ctx context.Context) Option { | ||||
| 	return func(o *Options) { | ||||
| @@ -132,6 +176,13 @@ func WithName(n string) Option { | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithMeter sets the meter | ||||
| func WithMeter(m meter.Meter) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Meter = m | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithTimeFunc sets the func to obtain current time | ||||
| func WithTimeFunc(fn func() time.Time) Option { | ||||
| 	return func(o *Options) { | ||||
| @@ -142,8 +193,8 @@ func WithTimeFunc(fn func() time.Time) Option { | ||||
| func WithZapKeys() Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.TimeKey = "@timestamp" | ||||
| 		o.LevelKey = "level" | ||||
| 		o.MessageKey = "msg" | ||||
| 		o.LevelKey = slog.LevelKey | ||||
| 		o.MessageKey = slog.MessageKey | ||||
| 		o.SourceKey = "caller" | ||||
| 		o.StacktraceKey = "stacktrace" | ||||
| 		o.ErrorKey = "error" | ||||
| @@ -152,8 +203,8 @@ func WithZapKeys() Option { | ||||
|  | ||||
| func WithZerologKeys() Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.TimeKey = "time" | ||||
| 		o.LevelKey = "level" | ||||
| 		o.TimeKey = slog.TimeKey | ||||
| 		o.LevelKey = slog.LevelKey | ||||
| 		o.MessageKey = "message" | ||||
| 		o.SourceKey = "caller" | ||||
| 		o.StacktraceKey = "stacktrace" | ||||
| @@ -175,10 +226,19 @@ func WithSlogKeys() Option { | ||||
| func WithMicroKeys() Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.TimeKey = "timestamp" | ||||
| 		o.LevelKey = "level" | ||||
| 		o.MessageKey = "msg" | ||||
| 		o.LevelKey = slog.LevelKey | ||||
| 		o.MessageKey = slog.MessageKey | ||||
| 		o.SourceKey = "caller" | ||||
| 		o.StacktraceKey = "stacktrace" | ||||
| 		o.ErrorKey = "error" | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithAddCallerSkipCount add skip count for copy logger | ||||
| func WithAddCallerSkipCount(n int) Option { | ||||
| 	return func(o *Options) { | ||||
| 		if n > 0 { | ||||
| 			o.CallerSkipCount += n | ||||
| 		} | ||||
| 	} | ||||
| } | ||||
|   | ||||
| @@ -2,18 +2,27 @@ package slog | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"fmt" | ||||
| 	"io" | ||||
| 	"log/slog" | ||||
| 	"os" | ||||
| 	"reflect" | ||||
| 	"regexp" | ||||
| 	"runtime" | ||||
| 	"strconv" | ||||
| 	"sync" | ||||
| 	"sync/atomic" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/logger" | ||||
| 	"go.unistack.org/micro/v3/semconv" | ||||
| 	"go.unistack.org/micro/v3/tracer" | ||||
| ) | ||||
|  | ||||
| const ( | ||||
| 	badKey = "!BADKEY" | ||||
| 	// defaultCallerSkipCount used by logger | ||||
| 	defaultCallerSkipCount = 3 | ||||
| 	timeFormat             = "2006-01-02T15:04:05.000000000Z07:00" | ||||
| ) | ||||
|  | ||||
| var reTrace = regexp.MustCompile(`.*/slog/logger\.go.*\n`) | ||||
|  | ||||
| var ( | ||||
| @@ -23,8 +32,30 @@ var ( | ||||
| 	warnValue  = slog.StringValue("warn") | ||||
| 	errorValue = slog.StringValue("error") | ||||
| 	fatalValue = slog.StringValue("fatal") | ||||
| 	noneValue  = slog.StringValue("none") | ||||
| ) | ||||
|  | ||||
| type wrapper struct { | ||||
| 	h     slog.Handler | ||||
| 	level atomic.Int64 | ||||
| } | ||||
|  | ||||
| func (h *wrapper) Enabled(ctx context.Context, level slog.Level) bool { | ||||
| 	return level >= slog.Level(int(h.level.Load())) | ||||
| } | ||||
|  | ||||
| func (h *wrapper) Handle(ctx context.Context, rec slog.Record) error { | ||||
| 	return h.h.Handle(ctx, rec) | ||||
| } | ||||
|  | ||||
| func (h *wrapper) WithAttrs(attrs []slog.Attr) slog.Handler { | ||||
| 	return h.h.WithAttrs(attrs) | ||||
| } | ||||
|  | ||||
| func (h *wrapper) WithGroup(name string) slog.Handler { | ||||
| 	return h.h.WithGroup(name) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) renameAttr(_ []string, a slog.Attr) slog.Attr { | ||||
| 	switch a.Key { | ||||
| 	case slog.SourceKey: | ||||
| @@ -33,6 +64,7 @@ func (s *slogLogger) renameAttr(_ []string, a slog.Attr) slog.Attr { | ||||
| 		a.Key = s.opts.SourceKey | ||||
| 	case slog.TimeKey: | ||||
| 		a.Key = s.opts.TimeKey | ||||
| 		a.Value = slog.StringValue(a.Value.Time().Format(timeFormat)) | ||||
| 	case slog.MessageKey: | ||||
| 		a.Key = s.opts.MessageKey | ||||
| 	case slog.LevelKey: | ||||
| @@ -52,6 +84,8 @@ func (s *slogLogger) renameAttr(_ []string, a slog.Attr) slog.Attr { | ||||
| 			a.Value = errorValue | ||||
| 		case lvl >= logger.FatalLevel: | ||||
| 			a.Value = fatalValue | ||||
| 		case lvl >= logger.NoneLevel: | ||||
| 			a.Value = noneValue | ||||
| 		default: | ||||
| 			a.Value = infoValue | ||||
| 		} | ||||
| @@ -61,8 +95,7 @@ func (s *slogLogger) renameAttr(_ []string, a slog.Attr) slog.Attr { | ||||
| } | ||||
|  | ||||
| type slogLogger struct { | ||||
| 	leveler *slog.LevelVar | ||||
| 	handler slog.Handler | ||||
| 	handler *wrapper | ||||
| 	opts    logger.Options | ||||
| 	mu      sync.RWMutex | ||||
| } | ||||
| @@ -76,51 +109,53 @@ func (s *slogLogger) Clone(opts ...logger.Option) logger.Logger { | ||||
| 		o(&options) | ||||
| 	} | ||||
|  | ||||
| 	l := &slogLogger{ | ||||
| 		opts: options, | ||||
| 	if len(options.ContextAttrFuncs) == 0 { | ||||
| 		options.ContextAttrFuncs = logger.DefaultContextAttrFuncs | ||||
| 	} | ||||
|  | ||||
| 	l.leveler = new(slog.LevelVar) | ||||
| 	handleOpt := &slog.HandlerOptions{ | ||||
| 		ReplaceAttr: l.renameAttr, | ||||
| 		Level:       l.leveler, | ||||
| 		AddSource:   l.opts.AddSource, | ||||
| 	attrs, _ := s.argsAttrs(options.Fields) | ||||
| 	l := &slogLogger{ | ||||
| 		handler: &wrapper{h: s.handler.h.WithAttrs(attrs)}, | ||||
| 		opts:    options, | ||||
| 	} | ||||
| 	l.leveler.Set(loggerToSlogLevel(l.opts.Level)) | ||||
| 	l.handler = slog.New(slog.NewJSONHandler(options.Out, handleOpt)).With(options.Fields...).Handler() | ||||
| 	l.handler.level.Store(int64(loggerToSlogLevel(options.Level))) | ||||
|  | ||||
| 	return l | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) V(level logger.Level) bool { | ||||
| 	return s.opts.Level.Enabled(level) | ||||
| 	s.mu.Lock() | ||||
| 	v := s.opts.Level.Enabled(level) | ||||
| 	s.mu.Unlock() | ||||
| 	return v | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Level(level logger.Level) { | ||||
| 	s.leveler.Set(loggerToSlogLevel(level)) | ||||
| 	s.mu.Lock() | ||||
| 	s.opts.Level = level | ||||
| 	s.handler.level.Store(int64(loggerToSlogLevel(level))) | ||||
| 	s.mu.Unlock() | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Options() logger.Options { | ||||
| 	return s.opts | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Fields(attrs ...interface{}) logger.Logger { | ||||
| func (s *slogLogger) Fields(fields ...interface{}) logger.Logger { | ||||
| 	s.mu.RLock() | ||||
| 	level := s.leveler.Level() | ||||
| 	options := s.opts | ||||
| 	s.mu.RUnlock() | ||||
|  | ||||
| 	l := &slogLogger{opts: options} | ||||
| 	l.leveler = new(slog.LevelVar) | ||||
| 	l.leveler.Set(level) | ||||
| 	logger.WithAddFields(fields...)(&l.opts) | ||||
|  | ||||
| 	handleOpt := &slog.HandlerOptions{ | ||||
| 		ReplaceAttr: l.renameAttr, | ||||
| 		Level:       l.leveler, | ||||
| 		AddSource:   l.opts.AddSource, | ||||
| 	if len(options.ContextAttrFuncs) == 0 { | ||||
| 		options.ContextAttrFuncs = logger.DefaultContextAttrFuncs | ||||
| 	} | ||||
|  | ||||
| 	l.handler = slog.New(slog.NewJSONHandler(l.opts.Out, handleOpt)).With(attrs...).Handler() | ||||
| 	attrs, _ := s.argsAttrs(fields) | ||||
| 	l.handler = &wrapper{h: s.handler.h.WithAttrs(attrs)} | ||||
| 	l.handler.level.Store(int64(loggerToSlogLevel(l.opts.Level))) | ||||
|  | ||||
| 	return l | ||||
| } | ||||
| @@ -128,393 +163,82 @@ func (s *slogLogger) Fields(attrs ...interface{}) logger.Logger { | ||||
| func (s *slogLogger) Init(opts ...logger.Option) error { | ||||
| 	s.mu.Lock() | ||||
|  | ||||
| 	if len(s.opts.ContextAttrFuncs) == 0 { | ||||
| 		s.opts.ContextAttrFuncs = logger.DefaultContextAttrFuncs | ||||
| 	} | ||||
|  | ||||
| 	for _, o := range opts { | ||||
| 		o(&s.opts) | ||||
| 	} | ||||
|  | ||||
| 	s.leveler = new(slog.LevelVar) | ||||
| 	if len(s.opts.ContextAttrFuncs) == 0 { | ||||
| 		s.opts.ContextAttrFuncs = logger.DefaultContextAttrFuncs | ||||
| 	} | ||||
|  | ||||
| 	handleOpt := &slog.HandlerOptions{ | ||||
| 		ReplaceAttr: s.renameAttr, | ||||
| 		Level:       s.leveler, | ||||
| 		Level:       loggerToSlogLevel(logger.TraceLevel), | ||||
| 		AddSource:   s.opts.AddSource, | ||||
| 	} | ||||
| 	s.leveler.Set(loggerToSlogLevel(s.opts.Level)) | ||||
| 	s.handler = slog.New(slog.NewJSONHandler(s.opts.Out, handleOpt)).With(s.opts.Fields...).Handler() | ||||
|  | ||||
| 	attrs, _ := s.argsAttrs(s.opts.Fields) | ||||
|  | ||||
| 	var h slog.Handler | ||||
| 	if s.opts.Context != nil { | ||||
| 		if v, ok := s.opts.Context.Value(handlerKey{}).(slog.Handler); ok && v != nil { | ||||
| 			h = v | ||||
| 		} | ||||
|  | ||||
| 		if fn := s.opts.Context.Value(handlerFnKey{}); fn != nil { | ||||
| 			if rfn := reflect.ValueOf(fn); rfn.Kind() == reflect.Func { | ||||
| 				if ret := rfn.Call([]reflect.Value{reflect.ValueOf(s.opts.Out), reflect.ValueOf(handleOpt)}); len(ret) == 1 { | ||||
| 					if iface, ok := ret[0].Interface().(slog.Handler); ok && iface != nil { | ||||
| 						h = iface | ||||
| 					} | ||||
| 				} | ||||
| 			} | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	if h == nil { | ||||
| 		h = slog.NewJSONHandler(s.opts.Out, handleOpt) | ||||
| 	} | ||||
|  | ||||
| 	s.handler = &wrapper{h: h.WithAttrs(attrs)} | ||||
| 	s.handler.level.Store(int64(loggerToSlogLevel(s.opts.Level))) | ||||
| 	s.mu.Unlock() | ||||
|  | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Log(ctx context.Context, lvl logger.Level, attrs ...interface{}) { | ||||
| 	if !s.V(lvl) { | ||||
| 		return | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), loggerToSlogLevel(lvl), fmt.Sprintf("%s", attrs[0]), pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	if s.opts.AddStacktrace && lvl == logger.ErrorLevel { | ||||
| 		stackInfo := make([]byte, 1024*1024) | ||||
| 		if stackSize := runtime.Stack(stackInfo, false); stackSize > 0 { | ||||
| 			traceLines := reTrace.Split(string(stackInfo[:stackSize]), -1) | ||||
| 			if len(traceLines) != 0 { | ||||
| 				attrs = append(attrs, slog.String(s.opts.StacktraceKey, traceLines[len(traceLines)-1])) | ||||
| 			} | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs[1:]...) | ||||
| 	r.Attrs(func(a slog.Attr) bool { | ||||
| 		if a.Key == s.opts.ErrorKey { | ||||
| 			if span, ok := tracer.SpanFromContext(ctx); ok { | ||||
| 				span.SetStatus(tracer.SpanStatusError, a.Value.String()) | ||||
| 				return false | ||||
| 			} | ||||
| 		} | ||||
| 		return true | ||||
| 	}) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| func (s *slogLogger) Log(ctx context.Context, lvl logger.Level, msg string, attrs ...interface{}) { | ||||
| 	s.printLog(ctx, lvl, msg, attrs...) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Logf(ctx context.Context, lvl logger.Level, msg string, attrs ...interface{}) { | ||||
| 	if !s.V(lvl) { | ||||
| 		return | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), loggerToSlogLevel(lvl), msg, pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	if s.opts.AddStacktrace && lvl == logger.ErrorLevel { | ||||
| 		stackInfo := make([]byte, 1024*1024) | ||||
| 		if stackSize := runtime.Stack(stackInfo, false); stackSize > 0 { | ||||
| 			traceLines := reTrace.Split(string(stackInfo[:stackSize]), -1) | ||||
| 			if len(traceLines) != 0 { | ||||
| 				attrs = append(attrs, (slog.String(s.opts.StacktraceKey, traceLines[len(traceLines)-1]))) | ||||
| 			} | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs[1:]...) | ||||
| 	r.Attrs(func(a slog.Attr) bool { | ||||
| 		if a.Key == s.opts.ErrorKey { | ||||
| 			if span, ok := tracer.SpanFromContext(ctx); ok { | ||||
| 				span.SetStatus(tracer.SpanStatusError, a.Value.String()) | ||||
| 				return false | ||||
| 			} | ||||
| 		} | ||||
| 		return true | ||||
| 	}) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| func (s *slogLogger) Info(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| 	s.printLog(ctx, logger.InfoLevel, msg, attrs...) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Info(ctx context.Context, attrs ...interface{}) { | ||||
| 	if !s.V(logger.InfoLevel) { | ||||
| 		return | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), slog.LevelInfo, fmt.Sprintf("%s", attrs[0]), pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs[1:]...) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| func (s *slogLogger) Debug(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| 	s.printLog(ctx, logger.DebugLevel, msg, attrs...) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Infof(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| 	if !s.V(logger.InfoLevel) { | ||||
| 		return | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), slog.LevelInfo, msg, pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs...) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| func (s *slogLogger) Trace(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| 	s.printLog(ctx, logger.TraceLevel, msg, attrs...) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Debug(ctx context.Context, attrs ...interface{}) { | ||||
| 	if !s.V(logger.DebugLevel) { | ||||
| 		return | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), slog.LevelDebug, fmt.Sprintf("%s", attrs[0]), pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs[1:]...) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| func (s *slogLogger) Error(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| 	s.printLog(ctx, logger.ErrorLevel, msg, attrs...) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Debugf(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| 	if !s.V(logger.DebugLevel) { | ||||
| 		return | ||||
| func (s *slogLogger) Fatal(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| 	s.printLog(ctx, logger.FatalLevel, msg, attrs...) | ||||
| 	for _, fn := range s.opts.FatalFinalizers { | ||||
| 		fn(ctx) | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), slog.LevelDebug, msg, pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	if closer, ok := s.opts.Out.(io.Closer); ok { | ||||
| 		closer.Close() | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs...) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Trace(ctx context.Context, attrs ...interface{}) { | ||||
| 	if !s.V(logger.TraceLevel) { | ||||
| 		return | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), slog.LevelDebug-1, fmt.Sprintf("%s", attrs[0]), pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs[1:]...) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Tracef(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| 	if !s.V(logger.TraceLevel) { | ||||
| 		return | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), slog.LevelDebug-1, msg, pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs[1:]...) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Error(ctx context.Context, attrs ...interface{}) { | ||||
| 	if !s.V(logger.ErrorLevel) { | ||||
| 		return | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), slog.LevelError, fmt.Sprintf("%s", attrs[0]), pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	if s.opts.AddStacktrace { | ||||
| 		stackInfo := make([]byte, 1024*1024) | ||||
| 		if stackSize := runtime.Stack(stackInfo, false); stackSize > 0 { | ||||
| 			traceLines := reTrace.Split(string(stackInfo[:stackSize]), -1) | ||||
| 			if len(traceLines) != 0 { | ||||
| 				attrs = append(attrs, slog.String("stacktrace", traceLines[len(traceLines)-1])) | ||||
| 			} | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs[1:]...) | ||||
| 	r.Attrs(func(a slog.Attr) bool { | ||||
| 		if a.Key == s.opts.ErrorKey { | ||||
| 			if span, ok := tracer.SpanFromContext(ctx); ok { | ||||
| 				span.SetStatus(tracer.SpanStatusError, a.Value.String()) | ||||
| 				return false | ||||
| 			} | ||||
| 		} | ||||
| 		return true | ||||
| 	}) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Errorf(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| 	if !s.V(logger.ErrorLevel) { | ||||
| 		return | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), slog.LevelError, msg, pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	if s.opts.AddStacktrace { | ||||
| 		stackInfo := make([]byte, 1024*1024) | ||||
| 		if stackSize := runtime.Stack(stackInfo, false); stackSize > 0 { | ||||
| 			traceLines := reTrace.Split(string(stackInfo[:stackSize]), -1) | ||||
| 			if len(traceLines) != 0 { | ||||
| 				attrs = append(attrs, slog.String("stacktrace", traceLines[len(traceLines)-1])) | ||||
| 			} | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs...) | ||||
| 	r.Attrs(func(a slog.Attr) bool { | ||||
| 		if a.Key == s.opts.ErrorKey { | ||||
| 			if span, ok := tracer.SpanFromContext(ctx); ok { | ||||
| 				span.SetStatus(tracer.SpanStatusError, a.Value.String()) | ||||
| 				return false | ||||
| 			} | ||||
| 		} | ||||
| 		return true | ||||
| 	}) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Fatal(ctx context.Context, attrs ...interface{}) { | ||||
| 	if !s.V(logger.FatalLevel) { | ||||
| 		return | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), slog.LevelError+1, fmt.Sprintf("%s", attrs[0]), pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs[1:]...) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| 	os.Exit(1) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Fatalf(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| 	if !s.V(logger.FatalLevel) { | ||||
| 		return | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), slog.LevelError+1, msg, pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs...) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| 	os.Exit(1) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Warn(ctx context.Context, attrs ...interface{}) { | ||||
| 	if !s.V(logger.WarnLevel) { | ||||
| 		return | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), slog.LevelWarn, fmt.Sprintf("%s", attrs[0]), pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs[1:]...) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Warnf(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| 	if !s.V(logger.WarnLevel) { | ||||
| 		return | ||||
| 	} | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, Infof] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), slog.LevelWarn, msg, pcs[0]) | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		attrs = append(attrs, fn(ctx)...) | ||||
| 	} | ||||
|  | ||||
| 	for idx, attr := range attrs { | ||||
| 		if ve, ok := attr.(error); ok && ve != nil { | ||||
| 			attrs[idx] = slog.String(s.opts.ErrorKey, ve.Error()) | ||||
| 			break | ||||
| 		} | ||||
| 	} | ||||
| 	r.Add(attrs[1:]...) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| func (s *slogLogger) Warn(ctx context.Context, msg string, attrs ...interface{}) { | ||||
| 	s.printLog(ctx, logger.WarnLevel, msg, attrs...) | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) Name() string { | ||||
| @@ -525,10 +249,59 @@ func (s *slogLogger) String() string { | ||||
| 	return "slog" | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) printLog(ctx context.Context, lvl logger.Level, msg string, args ...interface{}) { | ||||
| 	if !s.V(lvl) { | ||||
| 		return | ||||
| 	} | ||||
| 	var argError error | ||||
|  | ||||
| 	s.opts.Meter.Counter(semconv.LoggerMessageTotal, "level", lvl.String()).Inc() | ||||
|  | ||||
| 	attrs, err := s.argsAttrs(args) | ||||
| 	if err != nil { | ||||
| 		argError = err | ||||
| 	} | ||||
| 	if argError != nil { | ||||
| 		if span, ok := tracer.SpanFromContext(ctx); ok { | ||||
| 			span.SetStatus(tracer.SpanStatusError, argError.Error()) | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	for _, fn := range s.opts.ContextAttrFuncs { | ||||
| 		ctxAttrs, err := s.argsAttrs(fn(ctx)) | ||||
| 		if err != nil { | ||||
| 			argError = err | ||||
| 		} | ||||
| 		attrs = append(attrs, ctxAttrs...) | ||||
| 	} | ||||
| 	if argError != nil { | ||||
| 		if span, ok := tracer.SpanFromContext(ctx); ok { | ||||
| 			span.SetStatus(tracer.SpanStatusError, argError.Error()) | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	if s.opts.AddStacktrace && (lvl == logger.FatalLevel || lvl == logger.ErrorLevel) { | ||||
| 		stackInfo := make([]byte, 1024*1024) | ||||
| 		if stackSize := runtime.Stack(stackInfo, false); stackSize > 0 { | ||||
| 			traceLines := reTrace.Split(string(stackInfo[:stackSize]), -1) | ||||
| 			if len(traceLines) != 0 { | ||||
| 				attrs = append(attrs, slog.String(s.opts.StacktraceKey, traceLines[len(traceLines)-1])) | ||||
| 			} | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	var pcs [1]uintptr | ||||
| 	runtime.Callers(s.opts.CallerSkipCount, pcs[:]) // skip [Callers, printLog, LogLvlMethod] | ||||
| 	r := slog.NewRecord(s.opts.TimeFunc(), loggerToSlogLevel(lvl), msg, pcs[0]) | ||||
| 	r.AddAttrs(attrs...) | ||||
| 	_ = s.handler.Handle(ctx, r) | ||||
| } | ||||
|  | ||||
| func NewLogger(opts ...logger.Option) logger.Logger { | ||||
| 	s := &slogLogger{ | ||||
| 		opts: logger.NewOptions(opts...), | ||||
| 	} | ||||
| 	s.opts.CallerSkipCount = defaultCallerSkipCount | ||||
|  | ||||
| 	return s | ||||
| } | ||||
| @@ -545,6 +318,8 @@ func loggerToSlogLevel(level logger.Level) slog.Level { | ||||
| 		return slog.LevelDebug - 1 | ||||
| 	case logger.FatalLevel: | ||||
| 		return slog.LevelError + 1 | ||||
| 	case logger.NoneLevel: | ||||
| 		return slog.LevelError + 2 | ||||
| 	default: | ||||
| 		return slog.LevelInfo | ||||
| 	} | ||||
| @@ -562,7 +337,45 @@ func slogToLoggerLevel(level slog.Level) logger.Level { | ||||
| 		return logger.TraceLevel | ||||
| 	case slog.LevelError + 1: | ||||
| 		return logger.FatalLevel | ||||
| 	case slog.LevelError + 2: | ||||
| 		return logger.NoneLevel | ||||
| 	default: | ||||
| 		return logger.InfoLevel | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (s *slogLogger) argsAttrs(args []interface{}) ([]slog.Attr, error) { | ||||
| 	attrs := make([]slog.Attr, 0, len(args)) | ||||
| 	var err error | ||||
|  | ||||
| 	for idx := 0; idx < len(args); idx++ { | ||||
| 		switch arg := args[idx].(type) { | ||||
| 		case slog.Attr: | ||||
| 			attrs = append(attrs, arg) | ||||
| 		case string: | ||||
| 			if idx+1 < len(args) { | ||||
| 				attrs = append(attrs, slog.Any(arg, args[idx+1])) | ||||
| 				idx++ | ||||
| 			} else { | ||||
| 				attrs = append(attrs, slog.String(badKey, arg)) | ||||
| 			} | ||||
| 		case error: | ||||
| 			attrs = append(attrs, slog.String(s.opts.ErrorKey, arg.Error())) | ||||
| 			err = arg | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	return attrs, err | ||||
| } | ||||
|  | ||||
| type handlerKey struct{} | ||||
|  | ||||
| func WithHandler(h slog.Handler) logger.Option { | ||||
| 	return logger.SetOption(handlerKey{}, h) | ||||
| } | ||||
|  | ||||
| type handlerFnKey struct{} | ||||
|  | ||||
| func WithHandlerFunc(fn any) logger.Option { | ||||
| 	return logger.SetOption(handlerFnKey{}, fn) | ||||
| } | ||||
|   | ||||
| @@ -3,13 +3,244 @@ package slog | ||||
| import ( | ||||
| 	"bytes" | ||||
| 	"context" | ||||
| 	"errors" | ||||
| 	"fmt" | ||||
| 	"log" | ||||
| 	"log/slog" | ||||
| 	"strings" | ||||
| 	"testing" | ||||
| 	"time" | ||||
|  | ||||
| 	"github.com/google/uuid" | ||||
| 	"go.unistack.org/micro/v3/logger" | ||||
| 	"go.unistack.org/micro/v3/metadata" | ||||
| 	"go.unistack.org/micro/v3/util/buffer" | ||||
| ) | ||||
|  | ||||
| // always first to have proper check | ||||
| func TestStacktrace(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| 	l := NewLogger(logger.WithLevel(logger.DebugLevel), logger.WithOutput(buf), | ||||
| 		WithHandlerFunc(slog.NewTextHandler), | ||||
| 		logger.WithAddStacktrace(true), | ||||
| 	) | ||||
| 	if err := l.Init(logger.WithFields("key1", "val1")); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l.Error(ctx, "msg1", errors.New("err")) | ||||
|  | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`slog_test.go:32`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestNoneLevel(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| 	l := NewLogger(logger.WithLevel(logger.NoneLevel), logger.WithOutput(buf), | ||||
| 		WithHandlerFunc(slog.NewTextHandler), | ||||
| 		logger.WithAddStacktrace(true), | ||||
| 	) | ||||
| 	if err := l.Init(logger.WithFields("key1", "val1")); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l.Error(ctx, "msg1", errors.New("err")) | ||||
|  | ||||
| 	if buf.Len() != 0 { | ||||
| 		t.Fatalf("logger none level not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestDelayedBuffer(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| 	dbuf := buffer.NewDelayedBuffer(100, 100*time.Millisecond, buf) | ||||
| 	l := NewLogger(logger.WithLevel(logger.ErrorLevel), logger.WithOutput(dbuf), | ||||
| 		WithHandlerFunc(slog.NewTextHandler), | ||||
| 		logger.WithAddStacktrace(true), | ||||
| 	) | ||||
| 	if err := l.Init(logger.WithFields("key1", "val1")); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l.Error(ctx, "msg1", errors.New("err")) | ||||
| 	time.Sleep(120 * time.Millisecond) | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`key1=val1`)) { | ||||
| 		t.Fatalf("logger delayed buffer not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestTime(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| 	l := NewLogger(logger.WithLevel(logger.ErrorLevel), logger.WithOutput(buf), | ||||
| 		WithHandlerFunc(slog.NewTextHandler), | ||||
| 		logger.WithAddStacktrace(true), | ||||
| 		logger.WithTimeFunc(func() time.Time { | ||||
| 			return time.Unix(0, 0).UTC() | ||||
| 		}), | ||||
| 	) | ||||
| 	if err := l.Init(logger.WithFields("key1", "val1")); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l.Error(ctx, "msg1", errors.New("err")) | ||||
|  | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`timestamp=1970-01-01T00:00:00.000000000Z`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestWithFields(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| 	l := NewLogger(logger.WithLevel(logger.InfoLevel), logger.WithOutput(buf), | ||||
| 		WithHandlerFunc(slog.NewTextHandler), | ||||
| 		logger.WithDedupKeys(true), | ||||
| 	) | ||||
| 	if err := l.Init(logger.WithFields("key1", "val1")); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l.Info(ctx, "msg1") | ||||
|  | ||||
| 	l = l.Fields("key1", "val2") | ||||
|  | ||||
| 	l.Info(ctx, "msg2") | ||||
|  | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`msg=msg2 key1=val1`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestWithDedupKeysWithAddFields(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| 	l := NewLogger(logger.WithLevel(logger.InfoLevel), logger.WithOutput(buf), | ||||
| 		WithHandlerFunc(slog.NewTextHandler), | ||||
| 		logger.WithDedupKeys(true), | ||||
| 	) | ||||
| 	if err := l.Init(logger.WithFields("key1", "val1")); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l.Info(ctx, "msg1") | ||||
|  | ||||
| 	if err := l.Init(logger.WithAddFields("key2", "val2")); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l.Info(ctx, "msg2") | ||||
|  | ||||
| 	if err := l.Init(logger.WithAddFields("key2", "val3", "key1", "val4")); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l.Info(ctx, "msg3") | ||||
|  | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`msg=msg3 key1=val4 key2=val3`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestWithHandlerFunc(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| 	l := NewLogger(logger.WithLevel(logger.InfoLevel), logger.WithOutput(buf), | ||||
| 		WithHandlerFunc(slog.NewTextHandler), | ||||
| 	) | ||||
| 	if err := l.Init(); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l.Info(ctx, "msg1") | ||||
|  | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`msg=msg1`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestWithAddFields(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| 	l := NewLogger(logger.WithLevel(logger.InfoLevel), logger.WithOutput(buf)) | ||||
| 	if err := l.Init(); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l.Info(ctx, "msg1") | ||||
|  | ||||
| 	if err := l.Init(logger.WithAddFields("key1", "val1")); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	l.Info(ctx, "msg2") | ||||
|  | ||||
| 	if err := l.Init(logger.WithAddFields("key2", "val2")); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	l.Info(ctx, "msg3") | ||||
|  | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`"key1"`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`"key2"`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestMultipleFieldsWithLevel(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| 	l := NewLogger(logger.WithLevel(logger.InfoLevel), logger.WithOutput(buf)) | ||||
| 	if err := l.Init(); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l = l.Fields("key", "val") | ||||
|  | ||||
| 	l.Info(ctx, "msg1") | ||||
| 	nl := l.Clone(logger.WithLevel(logger.DebugLevel)) | ||||
| 	nl.Debug(ctx, "msg2") | ||||
| 	l.Debug(ctx, "msg3") | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`"key":"val"`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`"msg1"`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`"msg2"`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| 	if bytes.Contains(buf.Bytes(), []byte(`"msg3"`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestMultipleFields(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| 	l := NewLogger(logger.WithLevel(logger.InfoLevel), logger.WithOutput(buf)) | ||||
| 	if err := l.Init(); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l = l.Fields("key", "val") | ||||
|  | ||||
| 	l = l.Fields("key1", "val1") | ||||
|  | ||||
| 	l.Info(ctx, "msg") | ||||
|  | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`"key":"val"`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`"key1":"val1"`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestError(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| @@ -29,13 +260,22 @@ func TestError(t *testing.T) { | ||||
|  | ||||
| func TestErrorf(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
|  | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| 	l := NewLogger(logger.WithLevel(logger.ErrorLevel), logger.WithOutput(buf), logger.WithAddStacktrace(true)) | ||||
| 	if err := l.Init(); err != nil { | ||||
| 	if err := l.Init(logger.WithContextAttrFuncs(func(_ context.Context) []interface{} { | ||||
| 		return nil | ||||
| 	})); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l.Errorf(ctx, "message", fmt.Errorf("error message")) | ||||
| 	l.Log(ctx, logger.ErrorLevel, "message", errors.New("error msg")) | ||||
|  | ||||
| 	l.Log(ctx, logger.ErrorLevel, "", errors.New("error msg")) | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`"error":"error msg"`)) { | ||||
| 		t.Fatalf("logger error not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
|  | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`"stacktrace":"`)) { | ||||
| 		t.Fatalf("logger stacktrace not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| @@ -99,6 +339,11 @@ func TestFromContextWithFields(t *testing.T) { | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`"key":"val"`)) { | ||||
| 		t.Fatalf("logger fields not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
|  | ||||
| 	l.Info(ctx, "test", "uncorrected number attributes") | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte(`"!BADKEY":"uncorrected number attributes"`)) { | ||||
| 		t.Fatalf("logger fields not works, buf contains: %s", buf.Bytes()) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestClone(t *testing.T) { | ||||
| @@ -174,3 +419,75 @@ func TestLogger(t *testing.T) { | ||||
| 		t.Fatalf("logger warn, buf %s", buf.Bytes()) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func Test_WithContextAttrFunc(t *testing.T) { | ||||
| 	loggerContextAttrFuncs := []logger.ContextAttrFunc{ | ||||
| 		func(ctx context.Context) []interface{} { | ||||
| 			md, ok := metadata.FromOutgoingContext(ctx) | ||||
| 			if !ok { | ||||
| 				return nil | ||||
| 			} | ||||
| 			attrs := make([]interface{}, 0, 10) | ||||
| 			for k, v := range md { | ||||
| 				key := strings.ToLower(k) | ||||
| 				switch key { | ||||
| 				case "x-request-id", "phone", "external-Id", "source-service", "x-app-install-id", "client-id", "client-ip": | ||||
| 					attrs = append(attrs, key, v) | ||||
| 				} | ||||
| 			} | ||||
| 			return attrs | ||||
| 		}, | ||||
| 	} | ||||
|  | ||||
| 	logger.DefaultContextAttrFuncs = append(logger.DefaultContextAttrFuncs, loggerContextAttrFuncs...) | ||||
|  | ||||
| 	ctx := context.TODO() | ||||
| 	ctx = metadata.AppendOutgoingContext(ctx, "X-Request-Id", uuid.New().String(), | ||||
| 		"Source-Service", "Test-System") | ||||
|  | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| 	l := NewLogger(logger.WithLevel(logger.TraceLevel), logger.WithOutput(buf)) | ||||
| 	if err := l.Init(); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
|  | ||||
| 	l.Info(ctx, "test message") | ||||
| 	if !(bytes.Contains(buf.Bytes(), []byte(`"level":"info"`)) && bytes.Contains(buf.Bytes(), []byte(`"msg":"test message"`))) { | ||||
| 		t.Fatalf("logger info, buf %s", buf.Bytes()) | ||||
| 	} | ||||
| 	if !(bytes.Contains(buf.Bytes(), []byte(`"x-request-id":`))) { | ||||
| 		t.Fatalf("logger info, buf %s", buf.Bytes()) | ||||
| 	} | ||||
| 	if !(bytes.Contains(buf.Bytes(), []byte(`"source-service":"Test-System"`))) { | ||||
| 		t.Fatalf("logger info, buf %s", buf.Bytes()) | ||||
| 	} | ||||
| 	buf.Reset() | ||||
| 	omd, _ := metadata.FromOutgoingContext(ctx) | ||||
| 	l.Info(ctx, "test message1") | ||||
| 	omd.Set("Source-Service", "Test-System2") | ||||
| 	l.Info(ctx, "test message2") | ||||
|  | ||||
| 	// t.Logf("xxx %s", buf.Bytes()) | ||||
| } | ||||
|  | ||||
| func TestFatalFinalizers(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
| 	buf := bytes.NewBuffer(nil) | ||||
| 	l := NewLogger( | ||||
| 		logger.WithLevel(logger.TraceLevel), | ||||
| 		logger.WithOutput(buf), | ||||
| 	) | ||||
| 	if err := l.Init( | ||||
| 		logger.WithFatalFinalizers(func(ctx context.Context) { | ||||
| 			l.Info(ctx, "fatal finalizer") | ||||
| 		})); err != nil { | ||||
| 		t.Fatal(err) | ||||
| 	} | ||||
| 	l.Fatal(ctx, "info_msg1") | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte("fatal finalizer")) { | ||||
| 		t.Fatalf("logger dont have fatal message, buf %s", buf.Bytes()) | ||||
| 	} | ||||
| 	if !bytes.Contains(buf.Bytes(), []byte("info_msg1")) { | ||||
| 		t.Fatalf("logger dont have info_msg1 message, buf %s", buf.Bytes()) | ||||
| 	} | ||||
| } | ||||
|   | ||||
| @@ -8,7 +8,7 @@ import ( | ||||
| 	"strconv" | ||||
| 	"strings" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/codec" | ||||
| 	"go.unistack.org/micro/v4/codec" | ||||
| ) | ||||
|  | ||||
| const sf = "0-+# " | ||||
| @@ -36,14 +36,14 @@ var ( | ||||
| 	circularShortBytes = []byte("<shown>") | ||||
| 	invalidAngleBytes  = []byte("<invalid>") | ||||
| 	filteredBytes      = []byte("<filtered>") | ||||
| 	openBracketBytes   = []byte("[") | ||||
| 	closeBracketBytes  = []byte("]") | ||||
| 	percentBytes       = []byte("%") | ||||
| 	precisionBytes     = []byte(".") | ||||
| 	openAngleBytes     = []byte("<") | ||||
| 	closeAngleBytes    = []byte(">") | ||||
| 	openMapBytes       = []byte("{") | ||||
| 	closeMapBytes      = []byte("}") | ||||
| 	// openBracketBytes   = []byte("[") | ||||
| 	// closeBracketBytes  = []byte("]") | ||||
| 	percentBytes    = []byte("%") | ||||
| 	precisionBytes  = []byte(".") | ||||
| 	openAngleBytes  = []byte("<") | ||||
| 	closeAngleBytes = []byte(">") | ||||
| 	openMapBytes    = []byte("{") | ||||
| 	closeMapBytes   = []byte("}") | ||||
| ) | ||||
|  | ||||
| type protoMessage interface { | ||||
| @@ -52,13 +52,15 @@ type protoMessage interface { | ||||
| } | ||||
|  | ||||
| type Wrapper struct { | ||||
| 	val              interface{} | ||||
| 	s                fmt.State | ||||
| 	pointers         map[uintptr]int | ||||
| 	opts             *Options | ||||
| 	pointers map[uintptr]int | ||||
| 	takeMap  map[int]bool | ||||
|  | ||||
| 	val  interface{} | ||||
| 	s    fmt.State | ||||
| 	opts *Options | ||||
|  | ||||
| 	depth            int | ||||
| 	ignoreNextType   bool | ||||
| 	takeMap          map[int]bool | ||||
| 	protoWrapperType bool | ||||
| 	sqlWrapperType   bool | ||||
| } | ||||
|   | ||||
| @@ -5,7 +5,7 @@ import ( | ||||
| 	"strings" | ||||
| 	"testing" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/codec" | ||||
| 	"go.unistack.org/micro/v4/codec" | ||||
| ) | ||||
|  | ||||
| func TestUnwrap(t *testing.T) { | ||||
| @@ -82,12 +82,12 @@ func TestTagged(t *testing.T) { | ||||
| func TestTaggedNested(t *testing.T) { | ||||
| 	type val struct { | ||||
| 		key string `logger:"take"` | ||||
| 		val string `logger:"omit"` | ||||
| 		// val string `logger:"omit"` | ||||
| 		unk string | ||||
| 	} | ||||
| 	type str struct { | ||||
| 		key string `logger:"omit"` | ||||
| 		val *val   `logger:"take"` | ||||
| 		// key string `logger:"omit"` | ||||
| 		val *val `logger:"take"` | ||||
| 	} | ||||
|  | ||||
| 	var iface interface{} | ||||
|   | ||||
| @@ -1,399 +0,0 @@ | ||||
| // Package wrapper provides wrapper for Logger | ||||
| package wrapper | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"fmt" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/client" | ||||
| 	"go.unistack.org/micro/v3/logger" | ||||
| 	"go.unistack.org/micro/v3/server" | ||||
| ) | ||||
|  | ||||
| var ( | ||||
| 	// DefaultClientCallObserver called by wrapper in client Call | ||||
| 	DefaultClientCallObserver = func(ctx context.Context, req client.Request, rsp interface{}, opts []client.CallOption, err error) []string { | ||||
| 		labels := []string{"service", req.Service(), "endpoint", req.Endpoint()} | ||||
| 		if err != nil { | ||||
| 			labels = append(labels, "error", err.Error()) | ||||
| 		} | ||||
| 		return labels | ||||
| 	} | ||||
|  | ||||
| 	// DefaultClientStreamObserver called by wrapper in client Stream | ||||
| 	DefaultClientStreamObserver = func(ctx context.Context, req client.Request, opts []client.CallOption, stream client.Stream, err error) []string { | ||||
| 		labels := []string{"service", req.Service(), "endpoint", req.Endpoint()} | ||||
| 		if err != nil { | ||||
| 			labels = append(labels, "error", err.Error()) | ||||
| 		} | ||||
| 		return labels | ||||
| 	} | ||||
|  | ||||
| 	// DefaultClientPublishObserver called by wrapper in client Publish | ||||
| 	DefaultClientPublishObserver = func(ctx context.Context, msg client.Message, opts []client.PublishOption, err error) []string { | ||||
| 		labels := []string{"endpoint", msg.Topic()} | ||||
| 		if err != nil { | ||||
| 			labels = append(labels, "error", err.Error()) | ||||
| 		} | ||||
| 		return labels | ||||
| 	} | ||||
|  | ||||
| 	// DefaultServerHandlerObserver called by wrapper in server Handler | ||||
| 	DefaultServerHandlerObserver = func(ctx context.Context, req server.Request, rsp interface{}, err error) []string { | ||||
| 		labels := []string{"service", req.Service(), "endpoint", req.Endpoint()} | ||||
| 		if err != nil { | ||||
| 			labels = append(labels, "error", err.Error()) | ||||
| 		} | ||||
| 		return labels | ||||
| 	} | ||||
|  | ||||
| 	// DefaultServerSubscriberObserver called by wrapper in server Subscriber | ||||
| 	DefaultServerSubscriberObserver = func(ctx context.Context, msg server.Message, err error) []string { | ||||
| 		labels := []string{"endpoint", msg.Topic()} | ||||
| 		if err != nil { | ||||
| 			labels = append(labels, "error", err.Error()) | ||||
| 		} | ||||
| 		return labels | ||||
| 	} | ||||
|  | ||||
| 	// DefaultClientCallFuncObserver called by wrapper in client CallFunc | ||||
| 	DefaultClientCallFuncObserver = func(ctx context.Context, addr string, req client.Request, rsp interface{}, opts client.CallOptions, err error) []string { | ||||
| 		labels := []string{"service", req.Service(), "endpoint", req.Endpoint()} | ||||
| 		if err != nil { | ||||
| 			labels = append(labels, "error", err.Error()) | ||||
| 		} | ||||
| 		return labels | ||||
| 	} | ||||
|  | ||||
| 	// DefaultSkipEndpoints wrapper not called for this endpoints | ||||
| 	DefaultSkipEndpoints = []string{"Meter.Metrics", "Health.Live", "Health.Ready", "Health.Version"} | ||||
| ) | ||||
|  | ||||
| type lWrapper struct { | ||||
| 	client.Client | ||||
| 	serverHandler    server.HandlerFunc | ||||
| 	serverSubscriber server.SubscriberFunc | ||||
| 	clientCallFunc   client.CallFunc | ||||
| 	opts             Options | ||||
| } | ||||
|  | ||||
| type ( | ||||
| 	// ClientCallObserver func signature | ||||
| 	ClientCallObserver func(context.Context, client.Request, interface{}, []client.CallOption, error) []string | ||||
| 	// ClientStreamObserver func signature | ||||
| 	ClientStreamObserver func(context.Context, client.Request, []client.CallOption, client.Stream, error) []string | ||||
| 	// ClientPublishObserver func signature | ||||
| 	ClientPublishObserver func(context.Context, client.Message, []client.PublishOption, error) []string | ||||
| 	// ClientCallFuncObserver func signature | ||||
| 	ClientCallFuncObserver func(context.Context, string, client.Request, interface{}, client.CallOptions, error) []string | ||||
| 	// ServerHandlerObserver func signature | ||||
| 	ServerHandlerObserver func(context.Context, server.Request, interface{}, error) []string | ||||
| 	// ServerSubscriberObserver func signature | ||||
| 	ServerSubscriberObserver func(context.Context, server.Message, error) []string | ||||
| ) | ||||
|  | ||||
| // Options struct for wrapper | ||||
| type Options struct { | ||||
| 	// Logger that used for log | ||||
| 	Logger logger.Logger | ||||
| 	// ServerHandlerObservers funcs | ||||
| 	ServerHandlerObservers []ServerHandlerObserver | ||||
| 	// ServerSubscriberObservers funcs | ||||
| 	ServerSubscriberObservers []ServerSubscriberObserver | ||||
| 	// ClientCallObservers funcs | ||||
| 	ClientCallObservers []ClientCallObserver | ||||
| 	// ClientStreamObservers funcs | ||||
| 	ClientStreamObservers []ClientStreamObserver | ||||
| 	// ClientPublishObservers funcs | ||||
| 	ClientPublishObservers []ClientPublishObserver | ||||
| 	// ClientCallFuncObservers funcs | ||||
| 	ClientCallFuncObservers []ClientCallFuncObserver | ||||
| 	// SkipEndpoints | ||||
| 	SkipEndpoints []string | ||||
| 	// Level for logger | ||||
| 	Level logger.Level | ||||
| 	// Enabled flag | ||||
| 	Enabled bool | ||||
| } | ||||
|  | ||||
| // Option func signature | ||||
| type Option func(*Options) | ||||
|  | ||||
| // NewOptions creates Options from Option slice | ||||
| func NewOptions(opts ...Option) Options { | ||||
| 	options := Options{ | ||||
| 		Logger:                    logger.DefaultLogger, | ||||
| 		Level:                     logger.TraceLevel, | ||||
| 		ClientCallObservers:       []ClientCallObserver{DefaultClientCallObserver}, | ||||
| 		ClientStreamObservers:     []ClientStreamObserver{DefaultClientStreamObserver}, | ||||
| 		ClientPublishObservers:    []ClientPublishObserver{DefaultClientPublishObserver}, | ||||
| 		ClientCallFuncObservers:   []ClientCallFuncObserver{DefaultClientCallFuncObserver}, | ||||
| 		ServerHandlerObservers:    []ServerHandlerObserver{DefaultServerHandlerObserver}, | ||||
| 		ServerSubscriberObservers: []ServerSubscriberObserver{DefaultServerSubscriberObserver}, | ||||
| 		SkipEndpoints:             DefaultSkipEndpoints, | ||||
| 	} | ||||
|  | ||||
| 	for _, o := range opts { | ||||
| 		o(&options) | ||||
| 	} | ||||
|  | ||||
| 	return options | ||||
| } | ||||
|  | ||||
| // WithEnabled enable/diable flag | ||||
| func WithEnabled(b bool) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Enabled = b | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithLevel log level | ||||
| func WithLevel(l logger.Level) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Level = l | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithLogger logger | ||||
| func WithLogger(l logger.Logger) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Logger = l | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClientCallObservers funcs | ||||
| func WithClientCallObservers(ob ...ClientCallObserver) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.ClientCallObservers = ob | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClientStreamObservers funcs | ||||
| func WithClientStreamObservers(ob ...ClientStreamObserver) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.ClientStreamObservers = ob | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClientPublishObservers funcs | ||||
| func WithClientPublishObservers(ob ...ClientPublishObserver) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.ClientPublishObservers = ob | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithClientCallFuncObservers funcs | ||||
| func WithClientCallFuncObservers(ob ...ClientCallFuncObserver) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.ClientCallFuncObservers = ob | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithServerHandlerObservers funcs | ||||
| func WithServerHandlerObservers(ob ...ServerHandlerObserver) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.ServerHandlerObservers = ob | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // WithServerSubscriberObservers funcs | ||||
| func WithServerSubscriberObservers(ob ...ServerSubscriberObserver) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.ServerSubscriberObservers = ob | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // SkipEndpoins | ||||
| func SkipEndpoints(eps ...string) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.SkipEndpoints = append(o.SkipEndpoints, eps...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (l *lWrapper) Call(ctx context.Context, req client.Request, rsp interface{}, opts ...client.CallOption) error { | ||||
| 	err := l.Client.Call(ctx, req, rsp, opts...) | ||||
|  | ||||
| 	endpoint := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint()) | ||||
| 	for _, ep := range l.opts.SkipEndpoints { | ||||
| 		if ep == endpoint { | ||||
| 			return err | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	if !l.opts.Enabled { | ||||
| 		return err | ||||
| 	} | ||||
|  | ||||
| 	var labels []string | ||||
| 	for _, o := range l.opts.ClientCallObservers { | ||||
| 		labels = append(labels, o(ctx, req, rsp, opts, err)...) | ||||
| 	} | ||||
| 	l.opts.Logger.Fields(labels).Log(ctx, l.opts.Level) | ||||
|  | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| func (l *lWrapper) Stream(ctx context.Context, req client.Request, opts ...client.CallOption) (client.Stream, error) { | ||||
| 	stream, err := l.Client.Stream(ctx, req, opts...) | ||||
|  | ||||
| 	endpoint := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint()) | ||||
| 	for _, ep := range l.opts.SkipEndpoints { | ||||
| 		if ep == endpoint { | ||||
| 			return stream, err | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	if !l.opts.Enabled { | ||||
| 		return stream, err | ||||
| 	} | ||||
|  | ||||
| 	var labels []string | ||||
| 	for _, o := range l.opts.ClientStreamObservers { | ||||
| 		labels = append(labels, o(ctx, req, opts, stream, err)...) | ||||
| 	} | ||||
| 	l.opts.Logger.Fields(labels).Log(ctx, l.opts.Level) | ||||
|  | ||||
| 	return stream, err | ||||
| } | ||||
|  | ||||
| func (l *lWrapper) Publish(ctx context.Context, msg client.Message, opts ...client.PublishOption) error { | ||||
| 	err := l.Client.Publish(ctx, msg, opts...) | ||||
|  | ||||
| 	endpoint := msg.Topic() | ||||
| 	for _, ep := range l.opts.SkipEndpoints { | ||||
| 		if ep == endpoint { | ||||
| 			return err | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	if !l.opts.Enabled { | ||||
| 		return err | ||||
| 	} | ||||
|  | ||||
| 	var labels []string | ||||
| 	for _, o := range l.opts.ClientPublishObservers { | ||||
| 		labels = append(labels, o(ctx, msg, opts, err)...) | ||||
| 	} | ||||
| 	l.opts.Logger.Fields(labels).Log(ctx, l.opts.Level) | ||||
|  | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| func (l *lWrapper) ServerHandler(ctx context.Context, req server.Request, rsp interface{}) error { | ||||
| 	err := l.serverHandler(ctx, req, rsp) | ||||
|  | ||||
| 	endpoint := req.Endpoint() | ||||
| 	for _, ep := range l.opts.SkipEndpoints { | ||||
| 		if ep == endpoint { | ||||
| 			return err | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	if !l.opts.Enabled { | ||||
| 		return err | ||||
| 	} | ||||
|  | ||||
| 	var labels []string | ||||
| 	for _, o := range l.opts.ServerHandlerObservers { | ||||
| 		labels = append(labels, o(ctx, req, rsp, err)...) | ||||
| 	} | ||||
| 	l.opts.Logger.Fields(labels).Log(ctx, l.opts.Level) | ||||
|  | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| func (l *lWrapper) ServerSubscriber(ctx context.Context, msg server.Message) error { | ||||
| 	err := l.serverSubscriber(ctx, msg) | ||||
|  | ||||
| 	endpoint := msg.Topic() | ||||
| 	for _, ep := range l.opts.SkipEndpoints { | ||||
| 		if ep == endpoint { | ||||
| 			return err | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	if !l.opts.Enabled { | ||||
| 		return err | ||||
| 	} | ||||
|  | ||||
| 	var labels []string | ||||
| 	for _, o := range l.opts.ServerSubscriberObservers { | ||||
| 		labels = append(labels, o(ctx, msg, err)...) | ||||
| 	} | ||||
| 	l.opts.Logger.Fields(labels).Log(ctx, l.opts.Level) | ||||
|  | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| // NewClientWrapper accepts an open options and returns a Client Wrapper | ||||
| func NewClientWrapper(opts ...Option) client.Wrapper { | ||||
| 	return func(c client.Client) client.Client { | ||||
| 		options := NewOptions() | ||||
| 		for _, o := range opts { | ||||
| 			o(&options) | ||||
| 		} | ||||
| 		return &lWrapper{opts: options, Client: c} | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // NewClientCallWrapper accepts an options and returns a Call Wrapper | ||||
| func NewClientCallWrapper(opts ...Option) client.CallWrapper { | ||||
| 	return func(h client.CallFunc) client.CallFunc { | ||||
| 		options := NewOptions() | ||||
| 		for _, o := range opts { | ||||
| 			o(&options) | ||||
| 		} | ||||
|  | ||||
| 		l := &lWrapper{opts: options, clientCallFunc: h} | ||||
| 		return l.ClientCallFunc | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (l *lWrapper) ClientCallFunc(ctx context.Context, addr string, req client.Request, rsp interface{}, opts client.CallOptions) error { | ||||
| 	err := l.clientCallFunc(ctx, addr, req, rsp, opts) | ||||
|  | ||||
| 	endpoint := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint()) | ||||
| 	for _, ep := range l.opts.SkipEndpoints { | ||||
| 		if ep == endpoint { | ||||
| 			return err | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	if !l.opts.Enabled { | ||||
| 		return err | ||||
| 	} | ||||
|  | ||||
| 	var labels []string | ||||
| 	for _, o := range l.opts.ClientCallFuncObservers { | ||||
| 		labels = append(labels, o(ctx, addr, req, rsp, opts, err)...) | ||||
| 	} | ||||
| 	l.opts.Logger.Fields(labels).Log(ctx, l.opts.Level) | ||||
|  | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| // NewServerHandlerWrapper accepts an options and returns a Handler Wrapper | ||||
| func NewServerHandlerWrapper(opts ...Option) server.HandlerWrapper { | ||||
| 	return func(h server.HandlerFunc) server.HandlerFunc { | ||||
| 		options := NewOptions() | ||||
| 		for _, o := range opts { | ||||
| 			o(&options) | ||||
| 		} | ||||
|  | ||||
| 		l := &lWrapper{opts: options, serverHandler: h} | ||||
| 		return l.ServerHandler | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // NewServerSubscriberWrapper accepts an options and returns a Subscriber Wrapper | ||||
| func NewServerSubscriberWrapper(opts ...Option) server.SubscriberWrapper { | ||||
| 	return func(h server.SubscriberFunc) server.SubscriberFunc { | ||||
| 		options := NewOptions() | ||||
| 		for _, o := range opts { | ||||
| 			o(&options) | ||||
| 		} | ||||
|  | ||||
| 		l := &lWrapper{opts: options, serverSubscriber: h} | ||||
| 		return l.ServerSubscriber | ||||
| 	} | ||||
| } | ||||
| @@ -24,6 +24,17 @@ func FromIncomingContext(ctx context.Context) (Metadata, bool) { | ||||
| 	return md.md, ok | ||||
| } | ||||
|  | ||||
| // MustIncomingContext returns metadata from incoming ctx | ||||
| // returned metadata shoud not be modified or race condition happens. | ||||
| // If metadata not exists panics. | ||||
| func MustIncomingContext(ctx context.Context) Metadata { | ||||
| 	md, ok := FromIncomingContext(ctx) | ||||
| 	if !ok { | ||||
| 		panic("missing metadata") | ||||
| 	} | ||||
| 	return md | ||||
| } | ||||
|  | ||||
| // FromOutgoingContext returns metadata from outgoing ctx | ||||
| // returned metadata shoud not be modified or race condition happens | ||||
| func FromOutgoingContext(ctx context.Context) (Metadata, bool) { | ||||
| @@ -37,6 +48,17 @@ func FromOutgoingContext(ctx context.Context) (Metadata, bool) { | ||||
| 	return md.md, ok | ||||
| } | ||||
|  | ||||
| // MustOutgoingContext returns metadata from outgoing ctx | ||||
| // returned metadata shoud not be modified or race condition happens. | ||||
| // If metadata not exists panics. | ||||
| func MustOutgoingContext(ctx context.Context) Metadata { | ||||
| 	md, ok := FromOutgoingContext(ctx) | ||||
| 	if !ok { | ||||
| 		panic("missing metadata") | ||||
| 	} | ||||
| 	return md | ||||
| } | ||||
|  | ||||
| // FromContext returns metadata from the given context | ||||
| // returned metadata shoud not be modified or race condition happens | ||||
| func FromContext(ctx context.Context) (Metadata, bool) { | ||||
| @@ -50,15 +72,22 @@ func FromContext(ctx context.Context) (Metadata, bool) { | ||||
| 	return md.md, ok | ||||
| } | ||||
|  | ||||
| // MustContext returns metadata from the given context | ||||
| // returned metadata shoud not be modified or race condition happens | ||||
| func MustContext(ctx context.Context) Metadata { | ||||
| 	md, ok := FromContext(ctx) | ||||
| 	if !ok { | ||||
| 		panic("missing metadata") | ||||
| 	} | ||||
| 	return md | ||||
| } | ||||
|  | ||||
| // NewContext creates a new context with the given metadata | ||||
| func NewContext(ctx context.Context, md Metadata) context.Context { | ||||
| 	if ctx == nil { | ||||
| 		ctx = context.Background() | ||||
| 	} | ||||
| 	ctx = context.WithValue(ctx, mdKey{}, &rawMetadata{md}) | ||||
| 	ctx = context.WithValue(ctx, mdIncomingKey{}, &rawMetadata{}) | ||||
| 	ctx = context.WithValue(ctx, mdOutgoingKey{}, &rawMetadata{}) | ||||
| 	return ctx | ||||
| 	return context.WithValue(ctx, mdKey{}, &rawMetadata{md}) | ||||
| } | ||||
|  | ||||
| // SetOutgoingContext modify outgoing context with given metadata | ||||
| @@ -90,11 +119,7 @@ func NewIncomingContext(ctx context.Context, md Metadata) context.Context { | ||||
| 	if ctx == nil { | ||||
| 		ctx = context.Background() | ||||
| 	} | ||||
| 	ctx = context.WithValue(ctx, mdIncomingKey{}, &rawMetadata{md}) | ||||
| 	if v, ok := ctx.Value(mdOutgoingKey{}).(*rawMetadata); !ok || v == nil { | ||||
| 		ctx = context.WithValue(ctx, mdOutgoingKey{}, &rawMetadata{}) | ||||
| 	} | ||||
| 	return ctx | ||||
| 	return context.WithValue(ctx, mdIncomingKey{}, &rawMetadata{md}) | ||||
| } | ||||
|  | ||||
| // NewOutgoingContext creates a new context with outcoming metadata attached | ||||
| @@ -102,11 +127,7 @@ func NewOutgoingContext(ctx context.Context, md Metadata) context.Context { | ||||
| 	if ctx == nil { | ||||
| 		ctx = context.Background() | ||||
| 	} | ||||
| 	ctx = context.WithValue(ctx, mdOutgoingKey{}, &rawMetadata{md}) | ||||
| 	if v, ok := ctx.Value(mdIncomingKey{}).(*rawMetadata); !ok || v == nil { | ||||
| 		ctx = context.WithValue(ctx, mdIncomingKey{}, &rawMetadata{}) | ||||
| 	} | ||||
| 	return ctx | ||||
| 	return context.WithValue(ctx, mdOutgoingKey{}, &rawMetadata{md}) | ||||
| } | ||||
|  | ||||
| // AppendOutgoingContext apends new md to context | ||||
| @@ -122,7 +143,7 @@ func AppendOutgoingContext(ctx context.Context, kv ...string) context.Context { | ||||
| 	for k, v := range md { | ||||
| 		omd.Set(k, v) | ||||
| 	} | ||||
| 	return NewOutgoingContext(ctx, omd) | ||||
| 	return ctx | ||||
| } | ||||
|  | ||||
| // AppendIncomingContext apends new md to context | ||||
| @@ -138,5 +159,21 @@ func AppendIncomingContext(ctx context.Context, kv ...string) context.Context { | ||||
| 	for k, v := range md { | ||||
| 		omd.Set(k, v) | ||||
| 	} | ||||
| 	return NewIncomingContext(ctx, omd) | ||||
| 	return ctx | ||||
| } | ||||
|  | ||||
| // AppendContext apends new md to context | ||||
| func AppendContext(ctx context.Context, kv ...string) context.Context { | ||||
| 	md, ok := Pairs(kv...) | ||||
| 	if !ok { | ||||
| 		return ctx | ||||
| 	} | ||||
| 	omd, ok := FromContext(ctx) | ||||
| 	if !ok { | ||||
| 		return NewContext(ctx, md) | ||||
| 	} | ||||
| 	for k, v := range md { | ||||
| 		omd.Set(k, v) | ||||
| 	} | ||||
| 	return ctx | ||||
| } | ||||
|   | ||||
| @@ -1,9 +1,10 @@ | ||||
| // Package metadata is a way of defining message headers | ||||
| package metadata // import "go.unistack.org/micro/v3/metadata" | ||||
| package metadata | ||||
|  | ||||
| import ( | ||||
| 	"net/textproto" | ||||
| 	"sort" | ||||
| 	"strings" | ||||
| ) | ||||
|  | ||||
| var ( | ||||
| @@ -66,6 +67,14 @@ func (md Metadata) Iterator() *Iterator { | ||||
| 	return iter | ||||
| } | ||||
|  | ||||
| func (md Metadata) MustGet(key string) string { | ||||
| 	val, ok := md.Get(key) | ||||
| 	if !ok { | ||||
| 		panic("missing metadata key") | ||||
| 	} | ||||
| 	return val | ||||
| } | ||||
|  | ||||
| // Get returns value from metadata by key | ||||
| func (md Metadata) Get(key string) (string, bool) { | ||||
| 	// fast path | ||||
| @@ -73,6 +82,9 @@ func (md Metadata) Get(key string) (string, bool) { | ||||
| 	if !ok { | ||||
| 		// slow path | ||||
| 		val, ok = md[textproto.CanonicalMIMEHeaderKey(key)] | ||||
| 		if !ok { | ||||
| 			val, ok = md[strings.ToLower(key)] | ||||
| 		} | ||||
| 	} | ||||
| 	return val, ok | ||||
| } | ||||
| @@ -94,14 +106,23 @@ func (md Metadata) Del(keys ...string) { | ||||
| 		delete(md, key) | ||||
| 		// slow path | ||||
| 		delete(md, textproto.CanonicalMIMEHeaderKey(key)) | ||||
| 		// very slow path | ||||
| 		delete(md, strings.ToLower(key)) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Copy makes a copy of the metadata | ||||
| func (md Metadata) CopyTo(dst Metadata) { | ||||
| 	for k, v := range md { | ||||
| 		dst[k] = v | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Copy makes a copy of the metadata | ||||
| func Copy(md Metadata, exclude ...string) Metadata { | ||||
| 	nmd := New(len(md)) | ||||
| 	for key, val := range md { | ||||
| 		nmd.Set(key, val) | ||||
| 	for k, v := range md { | ||||
| 		nmd[k] = v | ||||
| 	} | ||||
| 	nmd.Del(exclude...) | ||||
| 	return nmd | ||||
| @@ -125,7 +146,7 @@ func Merge(omd Metadata, mmd Metadata, overwrite bool) Metadata { | ||||
| 		case ok && !overwrite: | ||||
| 			continue | ||||
| 		case val != "": | ||||
| 			nmd.Set(key, val) | ||||
| 			nmd[key] = val | ||||
| 		case ok && val == "": | ||||
| 			nmd.Del(key) | ||||
| 		} | ||||
| @@ -139,6 +160,8 @@ func Pairs(kv ...string) (Metadata, bool) { | ||||
| 		return nil, false | ||||
| 	} | ||||
| 	md := New(len(kv) / 2) | ||||
| 	md.Set(kv...) | ||||
| 	for idx := 0; idx < len(kv); idx += 2 { | ||||
| 		md[kv[idx]] = kv[idx+1] | ||||
| 	} | ||||
| 	return md, true | ||||
| } | ||||
|   | ||||
| @@ -5,6 +5,37 @@ import ( | ||||
| 	"testing" | ||||
| ) | ||||
|  | ||||
| func TestLowercase(t *testing.T) { | ||||
| 	md := New(1) | ||||
| 	md["x-request-id"] = "12345" | ||||
| 	v, ok := md.Get("X-Request-Id") | ||||
| 	if !ok || v == "" { | ||||
| 		t.Fatalf("metadata invalid %#+v", md) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestMultipleUsage(t *testing.T) { | ||||
| 	ctx := context.TODO() | ||||
| 	md := New(0) | ||||
| 	md.Set("key1_1", "val1_1", "key1_2", "val1_2", "key1_3", "val1_3") | ||||
| 	ctx = NewIncomingContext(ctx, Copy(md)) | ||||
| 	ctx = NewOutgoingContext(ctx, Copy(md)) | ||||
| 	imd, _ := FromIncomingContext(ctx) | ||||
| 	omd, _ := FromOutgoingContext(ctx) | ||||
| 	_ = func(x context.Context) context.Context { | ||||
| 		m, _ := FromIncomingContext(x) | ||||
| 		m.Del("key1_2") | ||||
| 		return ctx | ||||
| 	}(ctx) | ||||
| 	_ = func(x context.Context) context.Context { | ||||
| 		m, _ := FromIncomingContext(x) | ||||
| 		m.Del("key1_3") | ||||
| 		return ctx | ||||
| 	}(ctx) | ||||
| 	_ = imd | ||||
| 	_ = omd | ||||
| } | ||||
|  | ||||
| func TestMetadataSetMultiple(t *testing.T) { | ||||
| 	md := New(4) | ||||
| 	md.Set("key1", "val1", "key2", "val2", "key3") | ||||
| @@ -57,6 +88,13 @@ func TestPassing(t *testing.T) { | ||||
|  | ||||
| 	ctx = NewIncomingContext(ctx, md1) | ||||
| 	testCtx(ctx) | ||||
| 	_, ok := FromOutgoingContext(ctx) | ||||
| 	if ok { | ||||
| 		t.Fatalf("create outgoing context") | ||||
| 	} | ||||
|  | ||||
| 	ctx = NewOutgoingContext(ctx, New(1)) | ||||
| 	testCtx(ctx) | ||||
| 	md, ok := FromOutgoingContext(ctx) | ||||
| 	if !ok { | ||||
| 		t.Fatalf("missing metadata from outgoing context") | ||||
| @@ -80,7 +118,7 @@ func TestMerge(t *testing.T) { | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func TestIterator(t *testing.T) { | ||||
| func TestIterator(_ *testing.T) { | ||||
| 	md := Metadata{ | ||||
| 		"1Last":   "last", | ||||
| 		"2First":  "first", | ||||
|   | ||||
| @@ -15,6 +15,15 @@ func FromContext(ctx context.Context) (Meter, bool) { | ||||
| 	return c, ok | ||||
| } | ||||
|  | ||||
| // MustContext get meter from context | ||||
| func MustContext(ctx context.Context) Meter { | ||||
| 	m, ok := FromContext(ctx) | ||||
| 	if !ok { | ||||
| 		panic("missing meter") | ||||
| 	} | ||||
| 	return m | ||||
| } | ||||
|  | ||||
| // NewContext put meter in context | ||||
| func NewContext(ctx context.Context, c Meter) context.Context { | ||||
| 	if ctx == nil { | ||||
|   | ||||
| @@ -4,26 +4,31 @@ package meter | ||||
| import ( | ||||
| 	"io" | ||||
| 	"sort" | ||||
| 	"strconv" | ||||
| 	"strings" | ||||
| 	"sync" | ||||
| 	"time" | ||||
| ) | ||||
|  | ||||
| var ( | ||||
| 	// DefaultMeter is the default meter | ||||
| 	DefaultMeter Meter = NewMeter() | ||||
| 	DefaultMeter = NewMeter() | ||||
| 	// DefaultAddress data will be made available on this host:port | ||||
| 	DefaultAddress = ":9090" | ||||
| 	// DefaultPath the meter endpoint where the Meter data will be made available | ||||
| 	DefaultPath = "/metrics" | ||||
| 	// DefaultMetricPrefix holds the string that prepends to all metrics | ||||
| 	DefaultMetricPrefix = "micro_" | ||||
| 	// DefaultLabelPrefix holds the string that prepends to all labels | ||||
| 	DefaultLabelPrefix = "micro_" | ||||
| 	// DefaultMeterStatsInterval specifies interval for meter updating | ||||
| 	DefaultMeterStatsInterval = 5 * time.Second | ||||
| 	// DefaultSummaryQuantiles is the default spread of stats for summary | ||||
| 	DefaultSummaryQuantiles = []float64{0.5, 0.9, 0.97, 0.99, 1} | ||||
| 	// DefaultSummaryWindow is the default window for summary | ||||
| 	DefaultSummaryWindow = 5 * time.Minute | ||||
| 	// DefaultSkipEndpoints is the slice of endpoint that must not be metered | ||||
| 	DefaultSkipEndpoints = []string{ | ||||
| 		"MeterService.Metrics", | ||||
| 		"HealthService.Live", | ||||
| 		"HealthService.Ready", | ||||
| 		"HealthService.Version", | ||||
| 	} | ||||
| ) | ||||
|  | ||||
| // Meter is an interface for collecting and instrumenting metrics | ||||
| @@ -44,9 +49,11 @@ type Meter interface { | ||||
| 	Set(opts ...Option) Meter | ||||
| 	// Histogram get or create histogram | ||||
| 	Histogram(name string, labels ...string) Histogram | ||||
| 	// HistogramExt get or create histogram with specified quantiles | ||||
| 	HistogramExt(name string, quantiles []float64, labels ...string) Histogram | ||||
| 	// Summary get or create summary | ||||
| 	Summary(name string, labels ...string) Summary | ||||
| 	// SummaryExt get or create summary with spcified quantiles and window time | ||||
| 	// SummaryExt get or create summary with specified quantiles and window time | ||||
| 	SummaryExt(name string, window time.Duration, quantiles []float64, labels ...string) Summary | ||||
| 	// Write writes metrics to io.Writer | ||||
| 	Write(w io.Writer, opts ...Option) error | ||||
| @@ -54,6 +61,8 @@ type Meter interface { | ||||
| 	Options() Options | ||||
| 	// String return meter type | ||||
| 	String() string | ||||
| 	// Unregister metric name and drop all data | ||||
| 	Unregister(name string, labels ...string) bool | ||||
| } | ||||
|  | ||||
| // Counter is a counter | ||||
| @@ -75,7 +84,11 @@ type FloatCounter interface { | ||||
|  | ||||
| // Gauge is a float64 gauge | ||||
| type Gauge interface { | ||||
| 	Add(float64) | ||||
| 	Get() float64 | ||||
| 	Set(float64) | ||||
| 	Dec() | ||||
| 	Inc() | ||||
| } | ||||
|  | ||||
| // Histogram is a histogram for non-negative values with automatically created buckets | ||||
| @@ -112,6 +125,39 @@ func BuildLabels(labels ...string) []string { | ||||
| 	return labels | ||||
| } | ||||
|  | ||||
| var spool = newStringsPool(500) | ||||
|  | ||||
| type stringsPool struct { | ||||
| 	p *sync.Pool | ||||
| 	c int | ||||
| } | ||||
|  | ||||
| func newStringsPool(size int) *stringsPool { | ||||
| 	p := &stringsPool{c: size} | ||||
| 	p.p = &sync.Pool{ | ||||
| 		New: func() interface{} { | ||||
| 			return &strings.Builder{} | ||||
| 		}, | ||||
| 	} | ||||
| 	return p | ||||
| } | ||||
|  | ||||
| func (p *stringsPool) Cap() int { | ||||
| 	return p.c | ||||
| } | ||||
|  | ||||
| func (p *stringsPool) Get() *strings.Builder { | ||||
| 	return p.p.Get().(*strings.Builder) | ||||
| } | ||||
|  | ||||
| func (p *stringsPool) Put(b *strings.Builder) { | ||||
| 	if b.Cap() > p.c { | ||||
| 		return | ||||
| 	} | ||||
| 	b.Reset() | ||||
| 	p.p.Put(b) | ||||
| } | ||||
|  | ||||
| // BuildName used to combine metric with labels. | ||||
| // If labels count is odd, drop last element | ||||
| func BuildName(name string, labels ...string) string { | ||||
| @@ -120,8 +166,6 @@ func BuildName(name string, labels ...string) string { | ||||
| 	} | ||||
|  | ||||
| 	if len(labels) > 2 { | ||||
| 		sort.Sort(byKey(labels)) | ||||
|  | ||||
| 		idx := 0 | ||||
| 		for { | ||||
| 			if labels[idx] == labels[idx+2] { | ||||
| @@ -136,7 +180,9 @@ func BuildName(name string, labels ...string) string { | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	var b strings.Builder | ||||
| 	b := spool.Get() | ||||
| 	defer spool.Put(b) | ||||
|  | ||||
| 	_, _ = b.WriteString(name) | ||||
| 	_, _ = b.WriteRune('{') | ||||
| 	for idx := 0; idx < len(labels); idx += 2 { | ||||
| @@ -144,8 +190,9 @@ func BuildName(name string, labels ...string) string { | ||||
| 			_, _ = b.WriteRune(',') | ||||
| 		} | ||||
| 		_, _ = b.WriteString(labels[idx]) | ||||
| 		_, _ = b.WriteString(`=`) | ||||
| 		_, _ = b.WriteString(strconv.Quote(labels[idx+1])) | ||||
| 		_, _ = b.WriteString(`="`) | ||||
| 		_, _ = b.WriteString(labels[idx+1]) | ||||
| 		_, _ = b.WriteRune('"') | ||||
| 	} | ||||
| 	_, _ = b.WriteRune('}') | ||||
|  | ||||
|   | ||||
| @@ -50,11 +50,12 @@ func TestBuildName(t *testing.T) { | ||||
| 	data := map[string][]string{ | ||||
| 		`my_metric{firstlabel="value2",zerolabel="value3"}`: { | ||||
| 			"my_metric", | ||||
| 			"zerolabel", "value3", "firstlabel", "value2", | ||||
| 			"firstlabel", "value2", | ||||
| 			"zerolabel", "value3", | ||||
| 		}, | ||||
| 		`my_metric{broker="broker2",register="mdns",server="tcp"}`: { | ||||
| 			"my_metric", | ||||
| 			"broker", "broker1", "broker", "broker2", "server", "http", "server", "tcp", "register", "mdns", | ||||
| 			"broker", "broker1", "broker", "broker2", "register", "mdns", "server", "http", "server", "tcp", | ||||
| 		}, | ||||
| 		`my_metric{aaa="aaa"}`: { | ||||
| 			"my_metric", | ||||
|   | ||||
| @@ -28,6 +28,10 @@ func (r *noopMeter) Name() string { | ||||
| 	return r.opts.Name | ||||
| } | ||||
|  | ||||
| func (r *noopMeter) Unregister(name string, labels ...string) bool { | ||||
| 	return true | ||||
| } | ||||
|  | ||||
| // Init initialize options | ||||
| func (r *noopMeter) Init(opts ...Option) error { | ||||
| 	for _, o := range opts { | ||||
| @@ -37,32 +41,37 @@ func (r *noopMeter) Init(opts ...Option) error { | ||||
| } | ||||
|  | ||||
| // Counter implements the Meter interface | ||||
| func (r *noopMeter) Counter(name string, labels ...string) Counter { | ||||
| func (r *noopMeter) Counter(_ string, labels ...string) Counter { | ||||
| 	return &noopCounter{labels: labels} | ||||
| } | ||||
|  | ||||
| // FloatCounter implements the Meter interface | ||||
| func (r *noopMeter) FloatCounter(name string, labels ...string) FloatCounter { | ||||
| func (r *noopMeter) FloatCounter(_ string, labels ...string) FloatCounter { | ||||
| 	return &noopFloatCounter{labels: labels} | ||||
| } | ||||
|  | ||||
| // Gauge implements the Meter interface | ||||
| func (r *noopMeter) Gauge(name string, f func() float64, labels ...string) Gauge { | ||||
| func (r *noopMeter) Gauge(_ string, _ func() float64, labels ...string) Gauge { | ||||
| 	return &noopGauge{labels: labels} | ||||
| } | ||||
|  | ||||
| // Summary implements the Meter interface | ||||
| func (r *noopMeter) Summary(name string, labels ...string) Summary { | ||||
| func (r *noopMeter) Summary(_ string, labels ...string) Summary { | ||||
| 	return &noopSummary{labels: labels} | ||||
| } | ||||
|  | ||||
| // SummaryExt implements the Meter interface | ||||
| func (r *noopMeter) SummaryExt(name string, window time.Duration, quantiles []float64, labels ...string) Summary { | ||||
| func (r *noopMeter) SummaryExt(_ string, _ time.Duration, _ []float64, labels ...string) Summary { | ||||
| 	return &noopSummary{labels: labels} | ||||
| } | ||||
|  | ||||
| // Histogram implements the Meter interface | ||||
| func (r *noopMeter) Histogram(name string, labels ...string) Histogram { | ||||
| func (r *noopMeter) Histogram(_ string, labels ...string) Histogram { | ||||
| 	return &noopHistogram{labels: labels} | ||||
| } | ||||
|  | ||||
| // HistogramExt implements the Meter interface | ||||
| func (r *noopMeter) HistogramExt(_ string, quantiles []float64, labels ...string) Histogram { | ||||
| 	return &noopHistogram{labels: labels} | ||||
| } | ||||
|  | ||||
| @@ -77,7 +86,7 @@ func (r *noopMeter) Set(opts ...Option) Meter { | ||||
| 	return m | ||||
| } | ||||
|  | ||||
| func (r *noopMeter) Write(w io.Writer, opts ...Option) error { | ||||
| func (r *noopMeter) Write(_ io.Writer, _ ...Option) error { | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| @@ -132,6 +141,18 @@ type noopGauge struct { | ||||
| 	labels []string | ||||
| } | ||||
|  | ||||
| func (r *noopGauge) Add(float64) { | ||||
| } | ||||
|  | ||||
| func (r *noopGauge) Set(float64) { | ||||
| } | ||||
|  | ||||
| func (r *noopGauge) Inc() { | ||||
| } | ||||
|  | ||||
| func (r *noopGauge) Dec() { | ||||
| } | ||||
|  | ||||
| func (r *noopGauge) Get() float64 { | ||||
| 	return 0 | ||||
| } | ||||
|   | ||||
| @@ -2,17 +2,15 @@ package meter | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/logger" | ||||
| ) | ||||
|  | ||||
| var DefaultQuantiles = []float64{.005, .01, .025, .05, .1, .25, .5, 1, 2.5, 5, 10} | ||||
|  | ||||
| // Option powers the configuration for metrics implementations: | ||||
| type Option func(*Options) | ||||
|  | ||||
| // Options for metrics implementations | ||||
| type Options struct { | ||||
| 	// Logger used for logging | ||||
| 	Logger logger.Logger | ||||
| 	// Context holds external options | ||||
| 	Context context.Context | ||||
| 	// Name holds the meter name | ||||
| @@ -21,27 +19,22 @@ type Options struct { | ||||
| 	Address string | ||||
| 	// Path holds the path for metrics | ||||
| 	Path string | ||||
| 	// MetricPrefix holds the prefix for all metrics | ||||
| 	MetricPrefix string | ||||
| 	// LabelPrefix holds the prefix for all labels | ||||
| 	LabelPrefix string | ||||
| 	// Labels holds the default labels | ||||
| 	Labels []string | ||||
| 	// WriteProcessMetrics flag to write process metrics | ||||
| 	WriteProcessMetrics bool | ||||
| 	// WriteFDMetrics flag to write fd metrics | ||||
| 	WriteFDMetrics bool | ||||
| 	// Quantiles specifies buckets for histogram | ||||
| 	Quantiles []float64 | ||||
| } | ||||
|  | ||||
| // NewOptions prepares a set of options: | ||||
| func NewOptions(opt ...Option) Options { | ||||
| 	opts := Options{ | ||||
| 		Address:      DefaultAddress, | ||||
| 		Path:         DefaultPath, | ||||
| 		Context:      context.Background(), | ||||
| 		Logger:       logger.DefaultLogger, | ||||
| 		MetricPrefix: DefaultMetricPrefix, | ||||
| 		LabelPrefix:  DefaultLabelPrefix, | ||||
| 		Address: DefaultAddress, | ||||
| 		Path:    DefaultPath, | ||||
| 		Context: context.Background(), | ||||
| 	} | ||||
|  | ||||
| 	for _, o := range opt { | ||||
| @@ -51,20 +44,6 @@ func NewOptions(opt ...Option) Options { | ||||
| 	return opts | ||||
| } | ||||
|  | ||||
| // LabelPrefix sets the labels prefix | ||||
| func LabelPrefix(pref string) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.LabelPrefix = pref | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // MetricPrefix sets the metric prefix | ||||
| func MetricPrefix(pref string) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.MetricPrefix = pref | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Context sets the metrics context | ||||
| func Context(ctx context.Context) Option { | ||||
| 	return func(o *Options) { | ||||
| @@ -86,23 +65,14 @@ func Address(value string) Option { | ||||
| 	} | ||||
| } | ||||
|  | ||||
| /* | ||||
| // TimingObjectives defines the desired spread of statistics for histogram / timing metrics: | ||||
| func TimingObjectives(value map[float64]float64) Option { | ||||
| // Quantiles defines the desired spread of statistics for histogram metrics: | ||||
| func Quantiles(quantiles []float64) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.TimingObjectives = value | ||||
| 	} | ||||
| } | ||||
| */ | ||||
|  | ||||
| // Logger sets the logger | ||||
| func Logger(l logger.Logger) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Logger = l | ||||
| 		o.Quantiles = quantiles | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Labels sets the meter labels | ||||
| // Labels add the meter labels | ||||
| func Labels(ls ...string) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Labels = append(o.Labels, ls...) | ||||
|   | ||||
| @@ -1,347 +0,0 @@ | ||||
| package wrapper // import "go.unistack.org/micro/v3/meter/wrapper" | ||||
|  | ||||
| import ( | ||||
| 	"context" | ||||
| 	"fmt" | ||||
| 	"time" | ||||
|  | ||||
| 	"go.unistack.org/micro/v3/client" | ||||
| 	"go.unistack.org/micro/v3/meter" | ||||
| 	"go.unistack.org/micro/v3/server" | ||||
| ) | ||||
|  | ||||
| var ( | ||||
| 	// ClientRequestDurationSeconds specifies meter metric name | ||||
| 	ClientRequestDurationSeconds = "client_request_duration_seconds" | ||||
| 	// ClientRequestLatencyMicroseconds specifies meter metric name | ||||
| 	ClientRequestLatencyMicroseconds = "client_request_latency_microseconds" | ||||
| 	// ClientRequestTotal specifies meter metric name | ||||
| 	ClientRequestTotal = "client_request_total" | ||||
| 	// ClientRequestInflight specifies meter metric name | ||||
| 	ClientRequestInflight = "client_request_inflight" | ||||
| 	// ServerRequestDurationSeconds specifies meter metric name | ||||
| 	ServerRequestDurationSeconds = "server_request_duration_seconds" | ||||
| 	// ServerRequestLatencyMicroseconds specifies meter metric name | ||||
| 	ServerRequestLatencyMicroseconds = "server_request_latency_microseconds" | ||||
| 	// ServerRequestTotal specifies meter metric name | ||||
| 	ServerRequestTotal = "server_request_total" | ||||
| 	// ServerRequestInflight specifies meter metric name | ||||
| 	ServerRequestInflight = "server_request_inflight" | ||||
| 	// PublishMessageDurationSeconds specifies meter metric name | ||||
| 	PublishMessageDurationSeconds = "publish_message_duration_seconds" | ||||
| 	// PublishMessageLatencyMicroseconds specifies meter metric name | ||||
| 	PublishMessageLatencyMicroseconds = "publish_message_latency_microseconds" | ||||
| 	// PublishMessageTotal specifies meter metric name | ||||
| 	PublishMessageTotal = "publish_message_total" | ||||
| 	// PublishMessageInflight specifies meter metric name | ||||
| 	PublishMessageInflight = "publish_message_inflight" | ||||
| 	// SubscribeMessageDurationSeconds specifies meter metric name | ||||
| 	SubscribeMessageDurationSeconds = "subscribe_message_duration_seconds" | ||||
| 	// SubscribeMessageLatencyMicroseconds specifies meter metric name | ||||
| 	SubscribeMessageLatencyMicroseconds = "subscribe_message_latency_microseconds" | ||||
| 	// SubscribeMessageTotal specifies meter metric name | ||||
| 	SubscribeMessageTotal = "subscribe_message_total" | ||||
| 	// SubscribeMessageInflight specifies meter metric name | ||||
| 	SubscribeMessageInflight = "subscribe_message_inflight" | ||||
|  | ||||
| 	labelSuccess  = "success" | ||||
| 	labelFailure  = "failure" | ||||
| 	labelStatus   = "status" | ||||
| 	labelEndpoint = "endpoint" | ||||
|  | ||||
| 	// DefaultSkipEndpoints contains list of endpoints that not evaluted by wrapper | ||||
| 	DefaultSkipEndpoints = []string{"Meter.Metrics", "Health.Live", "Health.Ready", "Health.Version"} | ||||
| ) | ||||
|  | ||||
| // Options struct | ||||
| type Options struct { | ||||
| 	Meter         meter.Meter | ||||
| 	lopts         []meter.Option | ||||
| 	SkipEndpoints []string | ||||
| } | ||||
|  | ||||
| // Option func signature | ||||
| type Option func(*Options) | ||||
|  | ||||
| // NewOptions creates new Options struct | ||||
| func NewOptions(opts ...Option) Options { | ||||
| 	options := Options{ | ||||
| 		Meter:         meter.DefaultMeter, | ||||
| 		lopts:         make([]meter.Option, 0, 5), | ||||
| 		SkipEndpoints: DefaultSkipEndpoints, | ||||
| 	} | ||||
| 	for _, o := range opts { | ||||
| 		o(&options) | ||||
| 	} | ||||
| 	return options | ||||
| } | ||||
|  | ||||
| // ServiceName passes service name to meter label | ||||
| func ServiceName(name string) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.lopts = append(o.lopts, meter.Labels("name", name)) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // ServiceVersion passes service version to meter label | ||||
| func ServiceVersion(version string) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.lopts = append(o.lopts, meter.Labels("version", version)) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // ServiceID passes service id to meter label | ||||
| func ServiceID(id string) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.lopts = append(o.lopts, meter.Labels("id", id)) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // Meter passes meter | ||||
| func Meter(m meter.Meter) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.Meter = m | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // SkipEndoints add endpoint to skip | ||||
| func SkipEndoints(eps ...string) Option { | ||||
| 	return func(o *Options) { | ||||
| 		o.SkipEndpoints = append(o.SkipEndpoints, eps...) | ||||
| 	} | ||||
| } | ||||
|  | ||||
| type wrapper struct { | ||||
| 	client.Client | ||||
| 	callFunc client.CallFunc | ||||
| 	opts     Options | ||||
| } | ||||
|  | ||||
| // NewClientWrapper create new client wrapper | ||||
| func NewClientWrapper(opts ...Option) client.Wrapper { | ||||
| 	return func(c client.Client) client.Client { | ||||
| 		handler := &wrapper{ | ||||
| 			opts:   NewOptions(opts...), | ||||
| 			Client: c, | ||||
| 		} | ||||
| 		return handler | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // NewCallWrapper create new call wrapper | ||||
| func NewCallWrapper(opts ...Option) client.CallWrapper { | ||||
| 	return func(fn client.CallFunc) client.CallFunc { | ||||
| 		handler := &wrapper{ | ||||
| 			opts:     NewOptions(opts...), | ||||
| 			callFunc: fn, | ||||
| 		} | ||||
| 		return handler.CallFunc | ||||
| 	} | ||||
| } | ||||
|  | ||||
| func (w *wrapper) CallFunc(ctx context.Context, addr string, req client.Request, rsp interface{}, opts client.CallOptions) error { | ||||
| 	endpoint := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint()) | ||||
| 	for _, ep := range w.opts.SkipEndpoints { | ||||
| 		if ep == endpoint { | ||||
| 			return w.callFunc(ctx, addr, req, rsp, opts) | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	labels := make([]string, 0, 4) | ||||
| 	labels = append(labels, labelEndpoint, endpoint) | ||||
|  | ||||
| 	w.opts.Meter.Counter(ClientRequestInflight, labels...).Inc() | ||||
| 	ts := time.Now() | ||||
| 	err := w.callFunc(ctx, addr, req, rsp, opts) | ||||
| 	te := time.Since(ts) | ||||
| 	w.opts.Meter.Counter(ClientRequestInflight, labels...).Dec() | ||||
|  | ||||
| 	w.opts.Meter.Summary(ClientRequestLatencyMicroseconds, labels...).Update(te.Seconds()) | ||||
| 	w.opts.Meter.Histogram(ClientRequestDurationSeconds, labels...).Update(te.Seconds()) | ||||
|  | ||||
| 	if err == nil { | ||||
| 		labels = append(labels, labelStatus, labelSuccess) | ||||
| 	} else { | ||||
| 		labels = append(labels, labelStatus, labelFailure) | ||||
| 	} | ||||
| 	w.opts.Meter.Counter(ClientRequestTotal, labels...).Inc() | ||||
|  | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| func (w *wrapper) Call(ctx context.Context, req client.Request, rsp interface{}, opts ...client.CallOption) error { | ||||
| 	endpoint := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint()) | ||||
| 	for _, ep := range w.opts.SkipEndpoints { | ||||
| 		if ep == endpoint { | ||||
| 			return w.Client.Call(ctx, req, rsp, opts...) | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	labels := make([]string, 0, 4) | ||||
| 	labels = append(labels, labelEndpoint, endpoint) | ||||
|  | ||||
| 	w.opts.Meter.Counter(ClientRequestInflight, labels...).Inc() | ||||
| 	ts := time.Now() | ||||
| 	err := w.Client.Call(ctx, req, rsp, opts...) | ||||
| 	te := time.Since(ts) | ||||
| 	w.opts.Meter.Counter(ClientRequestInflight, labels...).Dec() | ||||
|  | ||||
| 	w.opts.Meter.Summary(ClientRequestLatencyMicroseconds, labels...).Update(te.Seconds()) | ||||
| 	w.opts.Meter.Histogram(ClientRequestDurationSeconds, labels...).Update(te.Seconds()) | ||||
|  | ||||
| 	if err == nil { | ||||
| 		labels = append(labels, labelStatus, labelSuccess) | ||||
| 	} else { | ||||
| 		labels = append(labels, labelStatus, labelFailure) | ||||
| 	} | ||||
| 	w.opts.Meter.Counter(ClientRequestTotal, labels...).Inc() | ||||
|  | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| func (w *wrapper) Stream(ctx context.Context, req client.Request, opts ...client.CallOption) (client.Stream, error) { | ||||
| 	endpoint := fmt.Sprintf("%s.%s", req.Service(), req.Endpoint()) | ||||
| 	for _, ep := range w.opts.SkipEndpoints { | ||||
| 		if ep == endpoint { | ||||
| 			return w.Client.Stream(ctx, req, opts...) | ||||
| 		} | ||||
| 	} | ||||
|  | ||||
| 	labels := make([]string, 0, 4) | ||||
| 	labels = append(labels, labelEndpoint, endpoint) | ||||
|  | ||||
| 	w.opts.Meter.Counter(ClientRequestInflight, labels...).Inc() | ||||
| 	ts := time.Now() | ||||
| 	stream, err := w.Client.Stream(ctx, req, opts...) | ||||
| 	te := time.Since(ts) | ||||
| 	w.opts.Meter.Counter(ClientRequestInflight, labels...).Dec() | ||||
|  | ||||
| 	w.opts.Meter.Summary(ClientRequestLatencyMicroseconds, labels...).Update(te.Seconds()) | ||||
| 	w.opts.Meter.Histogram(ClientRequestDurationSeconds, labels...).Update(te.Seconds()) | ||||
|  | ||||
| 	if err == nil { | ||||
| 		labels = append(labels, labelStatus, labelSuccess) | ||||
| 	} else { | ||||
| 		labels = append(labels, labelStatus, labelFailure) | ||||
| 	} | ||||
| 	w.opts.Meter.Counter(ClientRequestTotal, labels...).Inc() | ||||
|  | ||||
| 	return stream, err | ||||
| } | ||||
|  | ||||
| func (w *wrapper) Publish(ctx context.Context, p client.Message, opts ...client.PublishOption) error { | ||||
| 	endpoint := p.Topic() | ||||
|  | ||||
| 	labels := make([]string, 0, 4) | ||||
| 	labels = append(labels, labelEndpoint, endpoint) | ||||
|  | ||||
| 	w.opts.Meter.Counter(PublishMessageInflight, labels...).Inc() | ||||
| 	ts := time.Now() | ||||
| 	err := w.Client.Publish(ctx, p, opts...) | ||||
| 	te := time.Since(ts) | ||||
| 	w.opts.Meter.Counter(PublishMessageInflight, labels...).Dec() | ||||
|  | ||||
| 	w.opts.Meter.Summary(PublishMessageLatencyMicroseconds, labels...).Update(te.Seconds()) | ||||
| 	w.opts.Meter.Histogram(PublishMessageDurationSeconds, labels...).Update(te.Seconds()) | ||||
|  | ||||
| 	if err == nil { | ||||
| 		labels = append(labels, labelStatus, labelSuccess) | ||||
| 	} else { | ||||
| 		labels = append(labels, labelStatus, labelFailure) | ||||
| 	} | ||||
| 	w.opts.Meter.Counter(PublishMessageTotal, labels...).Inc() | ||||
|  | ||||
| 	return err | ||||
| } | ||||
|  | ||||
| // NewHandlerWrapper create new server handler wrapper | ||||
| // deprecated | ||||
| func NewHandlerWrapper(opts ...Option) server.HandlerWrapper { | ||||
| 	handler := &wrapper{ | ||||
| 		opts: NewOptions(opts...), | ||||
| 	} | ||||
| 	return handler.HandlerFunc | ||||
| } | ||||
|  | ||||
| // NewServerHandlerWrapper create new server handler wrapper | ||||
| func NewServerHandlerWrapper(opts ...Option) server.HandlerWrapper { | ||||
| 	handler := &wrapper{ | ||||
| 		opts: NewOptions(opts...), | ||||
| 	} | ||||
| 	return handler.HandlerFunc | ||||
| } | ||||
|  | ||||
| func (w *wrapper) HandlerFunc(fn server.HandlerFunc) server.HandlerFunc { | ||||
| 	return func(ctx context.Context, req server.Request, rsp interface{}) error { | ||||
| 		endpoint := req.Service() + "." + req.Endpoint() | ||||
| 		for _, ep := range w.opts.SkipEndpoints { | ||||
| 			if ep == endpoint { | ||||
| 				return fn(ctx, req, rsp) | ||||
| 			} | ||||
| 		} | ||||
|  | ||||
| 		labels := make([]string, 0, 4) | ||||
| 		labels = append(labels, labelEndpoint, endpoint) | ||||
|  | ||||
| 		w.opts.Meter.Counter(ServerRequestInflight, labels...).Inc() | ||||
| 		ts := time.Now() | ||||
| 		err := fn(ctx, req, rsp) | ||||
| 		te := time.Since(ts) | ||||
| 		w.opts.Meter.Counter(ServerRequestInflight, labels...).Dec() | ||||
|  | ||||
| 		w.opts.Meter.Summary(ServerRequestLatencyMicroseconds, labels...).Update(te.Seconds()) | ||||
| 		w.opts.Meter.Histogram(ServerRequestDurationSeconds, labels...).Update(te.Seconds()) | ||||
|  | ||||
| 		if err == nil { | ||||
| 			labels = append(labels, labelStatus, labelSuccess) | ||||
| 		} else { | ||||
| 			labels = append(labels, labelStatus, labelFailure) | ||||
| 		} | ||||
| 		w.opts.Meter.Counter(ServerRequestTotal, labels...).Inc() | ||||
|  | ||||
| 		return err | ||||
| 	} | ||||
| } | ||||
|  | ||||
| // NewSubscriberWrapper create server subscribe wrapper | ||||
| // deprecated | ||||
| func NewSubscriberWrapper(opts ...Option) server.SubscriberWrapper { | ||||
| 	handler := &wrapper{ | ||||
| 		opts: NewOptions(opts...), | ||||
| 	} | ||||
| 	return handler.SubscriberFunc | ||||
| } | ||||
|  | ||||
| func NewServerSubscriberWrapper(opts ...Option) server.SubscriberWrapper { | ||||
| 	handler := &wrapper{ | ||||
| 		opts: NewOptions(opts...), | ||||
| 	} | ||||
| 	return handler.SubscriberFunc | ||||
| } | ||||
|  | ||||
| func (w *wrapper) SubscriberFunc(fn server.SubscriberFunc) server.SubscriberFunc { | ||||
| 	return func(ctx context.Context, msg server.Message) error { | ||||
| 		endpoint := msg.Topic() | ||||
|  | ||||
| 		labels := make([]string, 0, 4) | ||||
| 		labels = append(labels, labelEndpoint, endpoint) | ||||
|  | ||||
| 		w.opts.Meter.Counter(SubscribeMessageInflight, labels...).Inc() | ||||
| 		ts := time.Now() | ||||
| 		err := fn(ctx, msg) | ||||
| 		te := time.Since(ts) | ||||
| 		w.opts.Meter.Counter(SubscribeMessageInflight, labels...).Dec() | ||||
|  | ||||
| 		w.opts.Meter.Summary(SubscribeMessageLatencyMicroseconds, labels...).Update(te.Seconds()) | ||||
| 		w.opts.Meter.Histogram(SubscribeMessageDurationSeconds, labels...).Update(te.Seconds()) | ||||
|  | ||||
| 		if err == nil { | ||||
| 			labels = append(labels, labelStatus, labelSuccess) | ||||
| 		} else { | ||||
| 			labels = append(labels, labelStatus, labelFailure) | ||||
| 		} | ||||
| 		w.opts.Meter.Counter(SubscribeMessageTotal, labels...).Inc() | ||||
|  | ||||
| 		return err | ||||
| 	} | ||||
| } | ||||
							
								
								
									
										36
									
								
								micro.go
									
									
									
									
									
								
							
							
						
						
									
										36
									
								
								micro.go
									
									
									
									
									
								
							| @@ -65,6 +65,8 @@ func As(b any, target any) bool { | ||||
| 			break | ||||
| 		case targetType.Implements(routerType): | ||||
| 			break | ||||
| 		case targetType.Implements(tracerType): | ||||
| 			break | ||||
| 		default: | ||||
| 			return false | ||||
| 		} | ||||
| @@ -76,19 +78,21 @@ func As(b any, target any) bool { | ||||
| 	return false | ||||
| } | ||||
|  | ||||
| var brokerType = reflect.TypeOf((*broker.Broker)(nil)).Elem() | ||||
| var loggerType = reflect.TypeOf((*logger.Logger)(nil)).Elem() | ||||
| var clientType = reflect.TypeOf((*client.Client)(nil)).Elem() | ||||
| var serverType = reflect.TypeOf((*server.Server)(nil)).Elem() | ||||
| var codecType = reflect.TypeOf((*codec.Codec)(nil)).Elem() | ||||
| var flowType = reflect.TypeOf((*flow.Flow)(nil)).Elem() | ||||
| var fsmType = reflect.TypeOf((*fsm.FSM)(nil)).Elem() | ||||
| var meterType = reflect.TypeOf((*meter.Meter)(nil)).Elem() | ||||
| var registerType = reflect.TypeOf((*register.Register)(nil)).Elem() | ||||
| var resolverType = reflect.TypeOf((*resolver.Resolver)(nil)).Elem() | ||||
| var routerType = reflect.TypeOf((*router.Router)(nil)).Elem() | ||||
| var selectorType = reflect.TypeOf((*selector.Selector)(nil)).Elem() | ||||
| var storeType = reflect.TypeOf((*store.Store)(nil)).Elem() | ||||
| var syncType = reflect.TypeOf((*sync.Sync)(nil)).Elem() | ||||
| var tracerType = reflect.TypeOf((*tracer.Tracer)(nil)).Elem() | ||||
| var serviceType = reflect.TypeOf((*Service)(nil)).Elem() | ||||
| var ( | ||||
| 	brokerType   = reflect.TypeOf((*broker.Broker)(nil)).Elem() | ||||
| 	loggerType   = reflect.TypeOf((*logger.Logger)(nil)).Elem() | ||||
| 	clientType   = reflect.TypeOf((*client.Client)(nil)).Elem() | ||||
| 	serverType   = reflect.TypeOf((*server.Server)(nil)).Elem() | ||||
| 	codecType    = reflect.TypeOf((*codec.Codec)(nil)).Elem() | ||||
| 	flowType     = reflect.TypeOf((*flow.Flow)(nil)).Elem() | ||||
| 	fsmType      = reflect.TypeOf((*fsm.FSM)(nil)).Elem() | ||||
| 	meterType    = reflect.TypeOf((*meter.Meter)(nil)).Elem() | ||||
| 	registerType = reflect.TypeOf((*register.Register)(nil)).Elem() | ||||
| 	resolverType = reflect.TypeOf((*resolver.Resolver)(nil)).Elem() | ||||
| 	routerType   = reflect.TypeOf((*router.Router)(nil)).Elem() | ||||
| 	selectorType = reflect.TypeOf((*selector.Selector)(nil)).Elem() | ||||
| 	storeType    = reflect.TypeOf((*store.Store)(nil)).Elem() | ||||
| 	syncType     = reflect.TypeOf((*sync.Sync)(nil)).Elem() | ||||
| 	tracerType   = reflect.TypeOf((*tracer.Tracer)(nil)).Elem() | ||||
| 	serviceType  = reflect.TypeOf((*Service)(nil)).Elem() | ||||
| ) | ||||
|   | ||||
| @@ -18,26 +18,27 @@ func TestAs(t *testing.T) { | ||||
| 	testCases := []struct { | ||||
| 		b      any | ||||
| 		target any | ||||
| 		match  bool | ||||
| 		want   any | ||||
|  | ||||
| 		match bool | ||||
| 	}{ | ||||
| 		{ | ||||
| 			broTarget, | ||||
| 			&b, | ||||
| 			true, | ||||
| 			broTarget, | ||||
| 			b:      broTarget, | ||||
| 			target: &b, | ||||
| 			match:  true, | ||||
| 			want:   broTarget, | ||||
| 		}, | ||||
| 		{ | ||||
| 			nil, | ||||
| 			&b, | ||||
| 			false, | ||||
| 			nil, | ||||
| 			b:      nil, | ||||
| 			target: &b, | ||||
| 			match:  false, | ||||
| 			want:   nil, | ||||
| 		}, | ||||
| 		{ | ||||
| 			fsmTarget, | ||||
| 			&b, | ||||
| 			false, | ||||
| 			nil, | ||||
| 			b:      fsmTarget, | ||||
| 			target: &b, | ||||
| 			match:  false, | ||||
| 			want:   nil, | ||||
| 		}, | ||||
| 	} | ||||
| 	for i, tc := range testCases { | ||||
| @@ -66,7 +67,13 @@ type bro struct { | ||||
|  | ||||
| func (p *bro) Name() string { return p.name } | ||||
|  | ||||
| func (p *bro) Init(opts ...broker.Option) error { return nil } | ||||
| func (p *bro) Live() bool { return true } | ||||
|  | ||||
| func (p *bro) Ready() bool { return true } | ||||
|  | ||||
| func (p *bro) Health() bool { return true } | ||||
|  | ||||
| func (p *bro) Init(_ ...broker.Option) error { return nil } | ||||
|  | ||||
| // Options returns broker options | ||||
| func (p *bro) Options() broker.Options { return broker.Options{} } | ||||
| @@ -75,28 +82,28 @@ func (p *bro) Options() broker.Options { return broker.Options{} } | ||||
| func (p *bro) Address() string { return "" } | ||||
|  | ||||
| // Connect connects to broker | ||||
| func (p *bro) Connect(ctx context.Context) error { return nil } | ||||
| func (p *bro) Connect(_ context.Context) error { return nil } | ||||
|  | ||||
| // Disconnect disconnect from broker | ||||
| func (p *bro) Disconnect(ctx context.Context) error { return nil } | ||||
| func (p *bro) Disconnect(_ context.Context) error { return nil } | ||||
|  | ||||
| // Publish message, msg can be single broker.Message or []broker.Message | ||||
| func (p *bro) Publish(ctx context.Context, topic string, msg *broker.Message, opts ...broker.PublishOption) error { | ||||
| func (p *bro) Publish(_ context.Context, _ string, _ *broker.Message, _ ...broker.PublishOption) error { | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| // BatchPublish messages to broker with multiple topics | ||||
| func (p *bro) BatchPublish(ctx context.Context, msgs []*broker.Message, opts ...broker.PublishOption) error { | ||||
| func (p *bro) BatchPublish(_ context.Context, _ []*broker.Message, _ ...broker.PublishOption) error { | ||||
| 	return nil | ||||
| } | ||||
|  | ||||
| // BatchSubscribe subscribes to topic messages via handler | ||||
| func (p *bro) BatchSubscribe(ctx context.Context, topic string, h broker.BatchHandler, opts ...broker.SubscribeOption) (broker.Subscriber, error) { | ||||
| func (p *bro) BatchSubscribe(_ context.Context, _ string, _ broker.BatchHandler, _ ...broker.SubscribeOption) (broker.Subscriber, error) { | ||||
| 	return nil, nil | ||||
| } | ||||
|  | ||||
| // Subscribe subscribes to topic message via handler | ||||
| func (p *bro) Subscribe(ctx context.Context, topic string, handler broker.Handler, opts ...broker.SubscribeOption) (broker.Subscriber, error) { | ||||
| func (p *bro) Subscribe(_ context.Context, _ string, _ broker.Handler, _ ...broker.SubscribeOption) (broker.Subscriber, error) { | ||||
| 	return nil, nil | ||||
| } | ||||
|  | ||||
| @@ -107,9 +114,9 @@ type fsmT struct { | ||||
| 	name string | ||||
| } | ||||
|  | ||||
| func (f *fsmT) Start(ctx context.Context, a interface{}, o ...Option) (interface{}, error) { | ||||
| func (f *fsmT) Start(_ context.Context, _ interface{}, _ ...Option) (interface{}, error) { | ||||
| 	return nil, nil | ||||
| } | ||||
| func (f *fsmT) Current() string                  { return f.name } | ||||
| func (f *fsmT) Reset()                           {} | ||||
| func (f *fsmT) State(s string, sf fsm.StateFunc) {} | ||||
| func (f *fsmT) Current() string                 { return f.name } | ||||
| func (f *fsmT) Reset()                          {} | ||||
| func (f *fsmT) State(_ string, _ fsm.StateFunc) {} | ||||
|   | ||||
| @@ -1,4 +1,4 @@ | ||||
| package mtls // import "go.unistack.org/micro/v3/mtls" | ||||
| package mtls | ||||
|  | ||||
| import ( | ||||
| 	"bytes" | ||||
|   | ||||
| @@ -8,19 +8,20 @@ import ( | ||||
|  | ||||
| // CertificateOptions holds options for x509.CreateCertificate | ||||
| type CertificateOptions struct { | ||||
| 	Organization          []string | ||||
| 	OrganizationalUnit    []string | ||||
| 	CommonName            string | ||||
| 	OCSPServer            []string | ||||
| 	IssuingCertificateURL []string | ||||
| 	SerialNumber          *big.Int | ||||
| 	NotAfter              time.Time | ||||
| 	NotBefore             time.Time | ||||
| 	SignatureAlgorithm    x509.SignatureAlgorithm | ||||
| 	PublicKeyAlgorithm    x509.PublicKeyAlgorithm | ||||
| 	CommonName            string | ||||
| 	Organization          []string | ||||
| 	OrganizationalUnit    []string | ||||
| 	OCSPServer            []string | ||||
| 	IssuingCertificateURL []string | ||||
| 	ExtKeyUsage           []x509.ExtKeyUsage | ||||
| 	KeyUsage              x509.KeyUsage | ||||
| 	IsCA                  bool | ||||
|  | ||||
| 	SignatureAlgorithm x509.SignatureAlgorithm | ||||
| 	PublicKeyAlgorithm x509.PublicKeyAlgorithm | ||||
| 	KeyUsage           x509.KeyUsage | ||||
| 	IsCA               bool | ||||
| } | ||||
|  | ||||
| // CertificateOrganizationalUnit set OrganizationalUnit in certificate subject | ||||
|   | ||||
| @@ -1,5 +1,5 @@ | ||||
| // Package network is for creating internetworks | ||||
| package network // import "go.unistack.org/micro/v3/network" | ||||
| package network | ||||
|  | ||||
| import ( | ||||
| 	"go.unistack.org/micro/v3/client" | ||||
|   | ||||
| @@ -119,7 +119,7 @@ func Tracer(t tracer.Tracer) Option { | ||||
| // NewOptions returns network default options | ||||
| func NewOptions(opts ...Option) Options { | ||||
| 	options := Options{ | ||||
| 		ID:      id.Must(), | ||||
| 		ID:      id.MustNew(), | ||||
| 		Name:    "go.micro", | ||||
| 		Address: ":0", | ||||
| 		Logger:  logger.DefaultLogger, | ||||
|   | ||||
Some files were not shown because too many files have changed in this diff Show More
		Reference in New Issue
	
	Block a user