mirror of
https://github.com/rocky-linux/peridot.git
synced 2024-12-26 04:00:55 +00:00
329 lines
11 KiB
Go
329 lines
11 KiB
Go
// Copyright 2016 Michal Witkowski. All Rights Reserved.
|
|
// See LICENSE for licensing terms.
|
|
|
|
package grpc_retry
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"io"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/grpc-ecosystem/go-grpc-middleware/util/metautils"
|
|
"golang.org/x/net/trace"
|
|
"google.golang.org/grpc"
|
|
"google.golang.org/grpc/codes"
|
|
"google.golang.org/grpc/metadata"
|
|
"google.golang.org/grpc/status"
|
|
)
|
|
|
|
const (
|
|
AttemptMetadataKey = "x-retry-attempty"
|
|
)
|
|
|
|
// UnaryClientInterceptor returns a new retrying unary client interceptor.
|
|
//
|
|
// The default configuration of the interceptor is to not retry *at all*. This behaviour can be
|
|
// changed through options (e.g. WithMax) on creation of the interceptor or on call (through grpc.CallOptions).
|
|
func UnaryClientInterceptor(optFuncs ...CallOption) grpc.UnaryClientInterceptor {
|
|
intOpts := reuseOrNewWithCallOptions(defaultOptions, optFuncs)
|
|
return func(parentCtx context.Context, method string, req, reply interface{}, cc *grpc.ClientConn, invoker grpc.UnaryInvoker, opts ...grpc.CallOption) error {
|
|
grpcOpts, retryOpts := filterCallOptions(opts)
|
|
callOpts := reuseOrNewWithCallOptions(intOpts, retryOpts)
|
|
// short circuit for simplicity, and avoiding allocations.
|
|
if callOpts.max == 0 {
|
|
return invoker(parentCtx, method, req, reply, cc, grpcOpts...)
|
|
}
|
|
var lastErr error
|
|
for attempt := uint(0); attempt < callOpts.max; attempt++ {
|
|
if err := waitRetryBackoff(attempt, parentCtx, callOpts); err != nil {
|
|
return err
|
|
}
|
|
callCtx := perCallContext(parentCtx, callOpts, attempt)
|
|
lastErr = invoker(callCtx, method, req, reply, cc, grpcOpts...)
|
|
// TODO(mwitkow): Maybe dial and transport errors should be retriable?
|
|
if lastErr == nil {
|
|
return nil
|
|
}
|
|
logTrace(parentCtx, "grpc_retry attempt: %d, got err: %v", attempt, lastErr)
|
|
if isContextError(lastErr) {
|
|
if parentCtx.Err() != nil {
|
|
logTrace(parentCtx, "grpc_retry attempt: %d, parent context error: %v", attempt, parentCtx.Err())
|
|
// its the parent context deadline or cancellation.
|
|
return lastErr
|
|
} else if callOpts.perCallTimeout != 0 {
|
|
// We have set a perCallTimeout in the retry middleware, which would result in a context error if
|
|
// the deadline was exceeded, in which case try again.
|
|
logTrace(parentCtx, "grpc_retry attempt: %d, context error from retry call", attempt)
|
|
continue
|
|
}
|
|
}
|
|
if !isRetriable(lastErr, callOpts) {
|
|
return lastErr
|
|
}
|
|
}
|
|
return lastErr
|
|
}
|
|
}
|
|
|
|
// StreamClientInterceptor returns a new retrying stream client interceptor for server side streaming calls.
|
|
//
|
|
// The default configuration of the interceptor is to not retry *at all*. This behaviour can be
|
|
// changed through options (e.g. WithMax) on creation of the interceptor or on call (through grpc.CallOptions).
|
|
//
|
|
// Retry logic is available *only for ServerStreams*, i.e. 1:n streams, as the internal logic needs
|
|
// to buffer the messages sent by the client. If retry is enabled on any other streams (ClientStreams,
|
|
// BidiStreams), the retry interceptor will fail the call.
|
|
func StreamClientInterceptor(optFuncs ...CallOption) grpc.StreamClientInterceptor {
|
|
intOpts := reuseOrNewWithCallOptions(defaultOptions, optFuncs)
|
|
return func(parentCtx context.Context, desc *grpc.StreamDesc, cc *grpc.ClientConn, method string, streamer grpc.Streamer, opts ...grpc.CallOption) (grpc.ClientStream, error) {
|
|
grpcOpts, retryOpts := filterCallOptions(opts)
|
|
callOpts := reuseOrNewWithCallOptions(intOpts, retryOpts)
|
|
// short circuit for simplicity, and avoiding allocations.
|
|
if callOpts.max == 0 {
|
|
return streamer(parentCtx, desc, cc, method, grpcOpts...)
|
|
}
|
|
if desc.ClientStreams {
|
|
return nil, status.Errorf(codes.Unimplemented, "grpc_retry: cannot retry on ClientStreams, set grpc_retry.Disable()")
|
|
}
|
|
|
|
var lastErr error
|
|
for attempt := uint(0); attempt < callOpts.max; attempt++ {
|
|
if err := waitRetryBackoff(attempt, parentCtx, callOpts); err != nil {
|
|
return nil, err
|
|
}
|
|
callCtx := perCallContext(parentCtx, callOpts, 0)
|
|
|
|
var newStreamer grpc.ClientStream
|
|
newStreamer, lastErr = streamer(callCtx, desc, cc, method, grpcOpts...)
|
|
if lastErr == nil {
|
|
retryingStreamer := &serverStreamingRetryingStream{
|
|
ClientStream: newStreamer,
|
|
callOpts: callOpts,
|
|
parentCtx: parentCtx,
|
|
streamerCall: func(ctx context.Context) (grpc.ClientStream, error) {
|
|
return streamer(ctx, desc, cc, method, grpcOpts...)
|
|
},
|
|
}
|
|
return retryingStreamer, nil
|
|
}
|
|
|
|
logTrace(parentCtx, "grpc_retry attempt: %d, got err: %v", attempt, lastErr)
|
|
if isContextError(lastErr) {
|
|
if parentCtx.Err() != nil {
|
|
logTrace(parentCtx, "grpc_retry attempt: %d, parent context error: %v", attempt, parentCtx.Err())
|
|
// its the parent context deadline or cancellation.
|
|
return nil, lastErr
|
|
} else if callOpts.perCallTimeout != 0 {
|
|
// We have set a perCallTimeout in the retry middleware, which would result in a context error if
|
|
// the deadline was exceeded, in which case try again.
|
|
logTrace(parentCtx, "grpc_retry attempt: %d, context error from retry call", attempt)
|
|
continue
|
|
}
|
|
}
|
|
if !isRetriable(lastErr, callOpts) {
|
|
return nil, lastErr
|
|
}
|
|
}
|
|
return nil, lastErr
|
|
}
|
|
}
|
|
|
|
// type serverStreamingRetryingStream is the implementation of grpc.ClientStream that acts as a
|
|
// proxy to the underlying call. If any of the RecvMsg() calls fail, it will try to reestablish
|
|
// a new ClientStream according to the retry policy.
|
|
type serverStreamingRetryingStream struct {
|
|
grpc.ClientStream
|
|
bufferedSends []interface{} // single message that the client can sen
|
|
receivedGood bool // indicates whether any prior receives were successful
|
|
wasClosedSend bool // indicates that CloseSend was closed
|
|
parentCtx context.Context
|
|
callOpts *options
|
|
streamerCall func(ctx context.Context) (grpc.ClientStream, error)
|
|
mu sync.RWMutex
|
|
}
|
|
|
|
func (s *serverStreamingRetryingStream) setStream(clientStream grpc.ClientStream) {
|
|
s.mu.Lock()
|
|
s.ClientStream = clientStream
|
|
s.mu.Unlock()
|
|
}
|
|
|
|
func (s *serverStreamingRetryingStream) getStream() grpc.ClientStream {
|
|
s.mu.RLock()
|
|
defer s.mu.RUnlock()
|
|
return s.ClientStream
|
|
}
|
|
|
|
func (s *serverStreamingRetryingStream) SendMsg(m interface{}) error {
|
|
s.mu.Lock()
|
|
s.bufferedSends = append(s.bufferedSends, m)
|
|
s.mu.Unlock()
|
|
return s.getStream().SendMsg(m)
|
|
}
|
|
|
|
func (s *serverStreamingRetryingStream) CloseSend() error {
|
|
s.mu.Lock()
|
|
s.wasClosedSend = true
|
|
s.mu.Unlock()
|
|
return s.getStream().CloseSend()
|
|
}
|
|
|
|
func (s *serverStreamingRetryingStream) Header() (metadata.MD, error) {
|
|
return s.getStream().Header()
|
|
}
|
|
|
|
func (s *serverStreamingRetryingStream) Trailer() metadata.MD {
|
|
return s.getStream().Trailer()
|
|
}
|
|
|
|
func (s *serverStreamingRetryingStream) RecvMsg(m interface{}) error {
|
|
attemptRetry, lastErr := s.receiveMsgAndIndicateRetry(m)
|
|
if !attemptRetry {
|
|
return lastErr // success or hard failure
|
|
}
|
|
// We start off from attempt 1, because zeroth was already made on normal SendMsg().
|
|
for attempt := uint(1); attempt < s.callOpts.max; attempt++ {
|
|
if err := waitRetryBackoff(attempt, s.parentCtx, s.callOpts); err != nil {
|
|
return err
|
|
}
|
|
callCtx := perCallContext(s.parentCtx, s.callOpts, attempt)
|
|
newStream, err := s.reestablishStreamAndResendBuffer(callCtx)
|
|
if err != nil {
|
|
// Retry dial and transport errors of establishing stream as grpc doesn't retry.
|
|
if isRetriable(err, s.callOpts) {
|
|
continue
|
|
}
|
|
return err
|
|
}
|
|
|
|
s.setStream(newStream)
|
|
attemptRetry, lastErr = s.receiveMsgAndIndicateRetry(m)
|
|
//fmt.Printf("Received message and indicate: %v %v\n", attemptRetry, lastErr)
|
|
if !attemptRetry {
|
|
return lastErr
|
|
}
|
|
}
|
|
return lastErr
|
|
}
|
|
|
|
func (s *serverStreamingRetryingStream) receiveMsgAndIndicateRetry(m interface{}) (bool, error) {
|
|
s.mu.RLock()
|
|
wasGood := s.receivedGood
|
|
s.mu.RUnlock()
|
|
err := s.getStream().RecvMsg(m)
|
|
if err == nil || err == io.EOF {
|
|
s.mu.Lock()
|
|
s.receivedGood = true
|
|
s.mu.Unlock()
|
|
return false, err
|
|
} else if wasGood {
|
|
// previous RecvMsg in the stream succeeded, no retry logic should interfere
|
|
return false, err
|
|
}
|
|
if isContextError(err) {
|
|
if s.parentCtx.Err() != nil {
|
|
logTrace(s.parentCtx, "grpc_retry parent context error: %v", s.parentCtx.Err())
|
|
return false, err
|
|
} else if s.callOpts.perCallTimeout != 0 {
|
|
// We have set a perCallTimeout in the retry middleware, which would result in a context error if
|
|
// the deadline was exceeded, in which case try again.
|
|
logTrace(s.parentCtx, "grpc_retry context error from retry call")
|
|
return true, err
|
|
}
|
|
}
|
|
return isRetriable(err, s.callOpts), err
|
|
}
|
|
|
|
func (s *serverStreamingRetryingStream) reestablishStreamAndResendBuffer(
|
|
callCtx context.Context,
|
|
) (grpc.ClientStream, error) {
|
|
s.mu.RLock()
|
|
bufferedSends := s.bufferedSends
|
|
s.mu.RUnlock()
|
|
newStream, err := s.streamerCall(callCtx)
|
|
if err != nil {
|
|
logTrace(callCtx, "grpc_retry failed redialing new stream: %v", err)
|
|
return nil, err
|
|
}
|
|
for _, msg := range bufferedSends {
|
|
if err := newStream.SendMsg(msg); err != nil {
|
|
logTrace(callCtx, "grpc_retry failed resending message: %v", err)
|
|
return nil, err
|
|
}
|
|
}
|
|
if err := newStream.CloseSend(); err != nil {
|
|
logTrace(callCtx, "grpc_retry failed CloseSend on new stream %v", err)
|
|
return nil, err
|
|
}
|
|
return newStream, nil
|
|
}
|
|
|
|
func waitRetryBackoff(attempt uint, parentCtx context.Context, callOpts *options) error {
|
|
var waitTime time.Duration = 0
|
|
if attempt > 0 {
|
|
waitTime = callOpts.backoffFunc(parentCtx, attempt)
|
|
}
|
|
if waitTime > 0 {
|
|
logTrace(parentCtx, "grpc_retry attempt: %d, backoff for %v", attempt, waitTime)
|
|
timer := time.NewTimer(waitTime)
|
|
select {
|
|
case <-parentCtx.Done():
|
|
timer.Stop()
|
|
return contextErrToGrpcErr(parentCtx.Err())
|
|
case <-timer.C:
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func isRetriable(err error, callOpts *options) bool {
|
|
errCode := status.Code(err)
|
|
if isContextError(err) {
|
|
// context errors are not retriable based on user settings.
|
|
return false
|
|
}
|
|
for _, code := range callOpts.codes {
|
|
if code == errCode {
|
|
return true
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
func isContextError(err error) bool {
|
|
code := status.Code(err)
|
|
return code == codes.DeadlineExceeded || code == codes.Canceled
|
|
}
|
|
|
|
func perCallContext(parentCtx context.Context, callOpts *options, attempt uint) context.Context {
|
|
ctx := parentCtx
|
|
if callOpts.perCallTimeout != 0 {
|
|
ctx, _ = context.WithTimeout(ctx, callOpts.perCallTimeout)
|
|
}
|
|
if attempt > 0 && callOpts.includeHeader {
|
|
mdClone := metautils.ExtractOutgoing(ctx).Clone().Set(AttemptMetadataKey, fmt.Sprintf("%d", attempt))
|
|
ctx = mdClone.ToOutgoing(ctx)
|
|
}
|
|
return ctx
|
|
}
|
|
|
|
func contextErrToGrpcErr(err error) error {
|
|
switch err {
|
|
case context.DeadlineExceeded:
|
|
return status.Error(codes.DeadlineExceeded, err.Error())
|
|
case context.Canceled:
|
|
return status.Error(codes.Canceled, err.Error())
|
|
default:
|
|
return status.Error(codes.Unknown, err.Error())
|
|
}
|
|
}
|
|
|
|
func logTrace(ctx context.Context, format string, a ...interface{}) {
|
|
tr, ok := trace.FromContext(ctx)
|
|
if !ok {
|
|
return
|
|
}
|
|
tr.LazyPrintf(format, a...)
|
|
}
|