// Copyright 2016 Michal Witkowski. All Rights Reserved. // See LICENSE for licensing terms. package grpc_retry import ( "context" "fmt" "io" "sync" "time" "github.com/grpc-ecosystem/go-grpc-middleware/util/metautils" "golang.org/x/net/trace" "google.golang.org/grpc" "google.golang.org/grpc/codes" "google.golang.org/grpc/metadata" "google.golang.org/grpc/status" ) const ( AttemptMetadataKey = "x-retry-attempty" ) // UnaryClientInterceptor returns a new retrying unary client interceptor. // // The default configuration of the interceptor is to not retry *at all*. This behaviour can be // changed through options (e.g. WithMax) on creation of the interceptor or on call (through grpc.CallOptions). func UnaryClientInterceptor(optFuncs ...CallOption) grpc.UnaryClientInterceptor { intOpts := reuseOrNewWithCallOptions(defaultOptions, optFuncs) return func(parentCtx context.Context, method string, req, reply interface{}, cc *grpc.ClientConn, invoker grpc.UnaryInvoker, opts ...grpc.CallOption) error { grpcOpts, retryOpts := filterCallOptions(opts) callOpts := reuseOrNewWithCallOptions(intOpts, retryOpts) // short circuit for simplicity, and avoiding allocations. if callOpts.max == 0 { return invoker(parentCtx, method, req, reply, cc, grpcOpts...) } var lastErr error for attempt := uint(0); attempt < callOpts.max; attempt++ { if err := waitRetryBackoff(attempt, parentCtx, callOpts); err != nil { return err } callCtx := perCallContext(parentCtx, callOpts, attempt) lastErr = invoker(callCtx, method, req, reply, cc, grpcOpts...) // TODO(mwitkow): Maybe dial and transport errors should be retriable? if lastErr == nil { return nil } logTrace(parentCtx, "grpc_retry attempt: %d, got err: %v", attempt, lastErr) if isContextError(lastErr) { if parentCtx.Err() != nil { logTrace(parentCtx, "grpc_retry attempt: %d, parent context error: %v", attempt, parentCtx.Err()) // its the parent context deadline or cancellation. return lastErr } else if callOpts.perCallTimeout != 0 { // We have set a perCallTimeout in the retry middleware, which would result in a context error if // the deadline was exceeded, in which case try again. logTrace(parentCtx, "grpc_retry attempt: %d, context error from retry call", attempt) continue } } if !isRetriable(lastErr, callOpts) { return lastErr } } return lastErr } } // StreamClientInterceptor returns a new retrying stream client interceptor for server side streaming calls. // // The default configuration of the interceptor is to not retry *at all*. This behaviour can be // changed through options (e.g. WithMax) on creation of the interceptor or on call (through grpc.CallOptions). // // Retry logic is available *only for ServerStreams*, i.e. 1:n streams, as the internal logic needs // to buffer the messages sent by the client. If retry is enabled on any other streams (ClientStreams, // BidiStreams), the retry interceptor will fail the call. func StreamClientInterceptor(optFuncs ...CallOption) grpc.StreamClientInterceptor { intOpts := reuseOrNewWithCallOptions(defaultOptions, optFuncs) return func(parentCtx context.Context, desc *grpc.StreamDesc, cc *grpc.ClientConn, method string, streamer grpc.Streamer, opts ...grpc.CallOption) (grpc.ClientStream, error) { grpcOpts, retryOpts := filterCallOptions(opts) callOpts := reuseOrNewWithCallOptions(intOpts, retryOpts) // short circuit for simplicity, and avoiding allocations. if callOpts.max == 0 { return streamer(parentCtx, desc, cc, method, grpcOpts...) } if desc.ClientStreams { return nil, status.Errorf(codes.Unimplemented, "grpc_retry: cannot retry on ClientStreams, set grpc_retry.Disable()") } var lastErr error for attempt := uint(0); attempt < callOpts.max; attempt++ { if err := waitRetryBackoff(attempt, parentCtx, callOpts); err != nil { return nil, err } callCtx := perCallContext(parentCtx, callOpts, 0) var newStreamer grpc.ClientStream newStreamer, lastErr = streamer(callCtx, desc, cc, method, grpcOpts...) if lastErr == nil { retryingStreamer := &serverStreamingRetryingStream{ ClientStream: newStreamer, callOpts: callOpts, parentCtx: parentCtx, streamerCall: func(ctx context.Context) (grpc.ClientStream, error) { return streamer(ctx, desc, cc, method, grpcOpts...) }, } return retryingStreamer, nil } logTrace(parentCtx, "grpc_retry attempt: %d, got err: %v", attempt, lastErr) if isContextError(lastErr) { if parentCtx.Err() != nil { logTrace(parentCtx, "grpc_retry attempt: %d, parent context error: %v", attempt, parentCtx.Err()) // its the parent context deadline or cancellation. return nil, lastErr } else if callOpts.perCallTimeout != 0 { // We have set a perCallTimeout in the retry middleware, which would result in a context error if // the deadline was exceeded, in which case try again. logTrace(parentCtx, "grpc_retry attempt: %d, context error from retry call", attempt) continue } } if !isRetriable(lastErr, callOpts) { return nil, lastErr } } return nil, lastErr } } // type serverStreamingRetryingStream is the implementation of grpc.ClientStream that acts as a // proxy to the underlying call. If any of the RecvMsg() calls fail, it will try to reestablish // a new ClientStream according to the retry policy. type serverStreamingRetryingStream struct { grpc.ClientStream bufferedSends []interface{} // single message that the client can sen receivedGood bool // indicates whether any prior receives were successful wasClosedSend bool // indicates that CloseSend was closed parentCtx context.Context callOpts *options streamerCall func(ctx context.Context) (grpc.ClientStream, error) mu sync.RWMutex } func (s *serverStreamingRetryingStream) setStream(clientStream grpc.ClientStream) { s.mu.Lock() s.ClientStream = clientStream s.mu.Unlock() } func (s *serverStreamingRetryingStream) getStream() grpc.ClientStream { s.mu.RLock() defer s.mu.RUnlock() return s.ClientStream } func (s *serverStreamingRetryingStream) SendMsg(m interface{}) error { s.mu.Lock() s.bufferedSends = append(s.bufferedSends, m) s.mu.Unlock() return s.getStream().SendMsg(m) } func (s *serverStreamingRetryingStream) CloseSend() error { s.mu.Lock() s.wasClosedSend = true s.mu.Unlock() return s.getStream().CloseSend() } func (s *serverStreamingRetryingStream) Header() (metadata.MD, error) { return s.getStream().Header() } func (s *serverStreamingRetryingStream) Trailer() metadata.MD { return s.getStream().Trailer() } func (s *serverStreamingRetryingStream) RecvMsg(m interface{}) error { attemptRetry, lastErr := s.receiveMsgAndIndicateRetry(m) if !attemptRetry { return lastErr // success or hard failure } // We start off from attempt 1, because zeroth was already made on normal SendMsg(). for attempt := uint(1); attempt < s.callOpts.max; attempt++ { if err := waitRetryBackoff(attempt, s.parentCtx, s.callOpts); err != nil { return err } callCtx := perCallContext(s.parentCtx, s.callOpts, attempt) newStream, err := s.reestablishStreamAndResendBuffer(callCtx) if err != nil { // Retry dial and transport errors of establishing stream as grpc doesn't retry. if isRetriable(err, s.callOpts) { continue } return err } s.setStream(newStream) attemptRetry, lastErr = s.receiveMsgAndIndicateRetry(m) //fmt.Printf("Received message and indicate: %v %v\n", attemptRetry, lastErr) if !attemptRetry { return lastErr } } return lastErr } func (s *serverStreamingRetryingStream) receiveMsgAndIndicateRetry(m interface{}) (bool, error) { s.mu.RLock() wasGood := s.receivedGood s.mu.RUnlock() err := s.getStream().RecvMsg(m) if err == nil || err == io.EOF { s.mu.Lock() s.receivedGood = true s.mu.Unlock() return false, err } else if wasGood { // previous RecvMsg in the stream succeeded, no retry logic should interfere return false, err } if isContextError(err) { if s.parentCtx.Err() != nil { logTrace(s.parentCtx, "grpc_retry parent context error: %v", s.parentCtx.Err()) return false, err } else if s.callOpts.perCallTimeout != 0 { // We have set a perCallTimeout in the retry middleware, which would result in a context error if // the deadline was exceeded, in which case try again. logTrace(s.parentCtx, "grpc_retry context error from retry call") return true, err } } return isRetriable(err, s.callOpts), err } func (s *serverStreamingRetryingStream) reestablishStreamAndResendBuffer(callCtx context.Context) (grpc.ClientStream, error) { s.mu.RLock() bufferedSends := s.bufferedSends s.mu.RUnlock() newStream, err := s.streamerCall(callCtx) if err != nil { logTrace(callCtx, "grpc_retry failed redialing new stream: %v", err) return nil, err } for _, msg := range bufferedSends { if err := newStream.SendMsg(msg); err != nil { logTrace(callCtx, "grpc_retry failed resending message: %v", err) return nil, err } } if err := newStream.CloseSend(); err != nil { logTrace(callCtx, "grpc_retry failed CloseSend on new stream %v", err) return nil, err } return newStream, nil } func waitRetryBackoff(attempt uint, parentCtx context.Context, callOpts *options) error { var waitTime time.Duration = 0 if attempt > 0 { waitTime = callOpts.backoffFunc(parentCtx, attempt) } if waitTime > 0 { logTrace(parentCtx, "grpc_retry attempt: %d, backoff for %v", attempt, waitTime) timer := time.NewTimer(waitTime) select { case <-parentCtx.Done(): timer.Stop() return contextErrToGrpcErr(parentCtx.Err()) case <-timer.C: } } return nil } func isRetriable(err error, callOpts *options) bool { errCode := status.Code(err) if isContextError(err) { // context errors are not retriable based on user settings. return false } for _, code := range callOpts.codes { if code == errCode { return true } } return false } func isContextError(err error) bool { code := status.Code(err) return code == codes.DeadlineExceeded || code == codes.Canceled } func perCallContext(parentCtx context.Context, callOpts *options, attempt uint) context.Context { ctx := parentCtx if callOpts.perCallTimeout != 0 { ctx, _ = context.WithTimeout(ctx, callOpts.perCallTimeout) } if attempt > 0 && callOpts.includeHeader { mdClone := metautils.ExtractOutgoing(ctx).Clone().Set(AttemptMetadataKey, fmt.Sprintf("%d", attempt)) ctx = mdClone.ToOutgoing(ctx) } return ctx } func contextErrToGrpcErr(err error) error { switch err { case context.DeadlineExceeded: return status.Errorf(codes.DeadlineExceeded, err.Error()) case context.Canceled: return status.Errorf(codes.Canceled, err.Error()) default: return status.Errorf(codes.Unknown, err.Error()) } } func logTrace(ctx context.Context, format string, a ...interface{}) { tr, ok := trace.FromContext(ctx) if !ok { return } tr.LazyPrintf(format, a...) }