mirror of
https://github.com/rocky-linux/peridot.git
synced 2024-12-26 20:20:56 +00:00
1569 lines
51 KiB
Go
1569 lines
51 KiB
Go
|
// The MIT License
|
||
|
//
|
||
|
// Copyright (c) 2020 Temporal Technologies Inc. All rights reserved.
|
||
|
//
|
||
|
// Copyright (c) 2020 Uber Technologies, Inc.
|
||
|
//
|
||
|
// Permission is hereby granted, free of charge, to any person obtaining a copy
|
||
|
// of this software and associated documentation files (the "Software"), to deal
|
||
|
// in the Software without restriction, including without limitation the rights
|
||
|
// to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
||
|
// copies of the Software, and to permit persons to whom the Software is
|
||
|
// furnished to do so, subject to the following conditions:
|
||
|
//
|
||
|
// The above copyright notice and this permission notice shall be included in
|
||
|
// all copies or substantial portions of the Software.
|
||
|
//
|
||
|
// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||
|
// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||
|
// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
||
|
// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||
|
// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
||
|
// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
||
|
// THE SOFTWARE.
|
||
|
|
||
|
package internal
|
||
|
|
||
|
// All code in this file is private to the package.
|
||
|
|
||
|
import (
|
||
|
"context"
|
||
|
"crypto/md5"
|
||
|
"encoding/hex"
|
||
|
"errors"
|
||
|
"fmt"
|
||
|
"io"
|
||
|
"math"
|
||
|
"os"
|
||
|
"reflect"
|
||
|
"runtime"
|
||
|
"strconv"
|
||
|
"strings"
|
||
|
"sync"
|
||
|
"time"
|
||
|
|
||
|
"github.com/gogo/protobuf/jsonpb"
|
||
|
"github.com/gogo/protobuf/proto"
|
||
|
"github.com/golang/mock/gomock"
|
||
|
"github.com/pborman/uuid"
|
||
|
commonpb "go.temporal.io/api/common/v1"
|
||
|
enumspb "go.temporal.io/api/enums/v1"
|
||
|
historypb "go.temporal.io/api/history/v1"
|
||
|
"go.temporal.io/api/workflowservice/v1"
|
||
|
"go.temporal.io/api/workflowservicemock/v1"
|
||
|
|
||
|
"go.temporal.io/sdk/converter"
|
||
|
"go.temporal.io/sdk/internal/common/metrics"
|
||
|
"go.temporal.io/sdk/internal/common/serializer"
|
||
|
"go.temporal.io/sdk/internal/common/util"
|
||
|
ilog "go.temporal.io/sdk/internal/log"
|
||
|
"go.temporal.io/sdk/log"
|
||
|
)
|
||
|
|
||
|
const (
|
||
|
// Set to 2 pollers for now, can adjust later if needed. The typical RTT (round-trip time) is below 1ms within data
|
||
|
// center. And the poll API latency is about 5ms. With 2 poller, we could achieve around 300~400 RPS.
|
||
|
defaultConcurrentPollRoutineSize = 2
|
||
|
|
||
|
defaultMaxConcurrentActivityExecutionSize = 1000 // Large concurrent activity execution size (1k)
|
||
|
defaultWorkerActivitiesPerSecond = 100000 // Large activity executions/sec (unlimited)
|
||
|
|
||
|
defaultMaxConcurrentLocalActivityExecutionSize = 1000 // Large concurrent activity execution size (1k)
|
||
|
defaultWorkerLocalActivitiesPerSecond = 100000 // Large activity executions/sec (unlimited)
|
||
|
|
||
|
defaultTaskQueueActivitiesPerSecond = 100000.0 // Large activity executions/sec (unlimited)
|
||
|
|
||
|
defaultMaxConcurrentTaskExecutionSize = 1000 // hardcoded max task execution size.
|
||
|
defaultWorkerTaskExecutionRate = 100000 // Large task execution rate (unlimited)
|
||
|
|
||
|
defaultPollerRate = 1000
|
||
|
|
||
|
defaultMaxConcurrentSessionExecutionSize = 1000 // Large concurrent session execution size (1k)
|
||
|
|
||
|
defaultDeadlockDetectionTimeout = time.Second // By default kill workflow tasks that are running more than 1 sec.
|
||
|
// Unlimited deadlock detection timeout is used when we want to allow workflow tasks to run indefinitely, such
|
||
|
// as during debugging.
|
||
|
unlimitedDeadlockDetectionTimeout = math.MaxInt64
|
||
|
|
||
|
testTagsContextKey = "temporal-testTags"
|
||
|
)
|
||
|
|
||
|
type (
|
||
|
// WorkflowWorker wraps the code for hosting workflow types.
|
||
|
// And worker is mapped 1:1 with task queue. If the user want's to poll multiple
|
||
|
// task queue names they might have to manage 'n' workers for 'n' task queues.
|
||
|
workflowWorker struct {
|
||
|
executionParameters workerExecutionParameters
|
||
|
workflowService workflowservice.WorkflowServiceClient
|
||
|
poller taskPoller // taskPoller to poll and process the tasks.
|
||
|
worker *baseWorker
|
||
|
localActivityWorker *baseWorker
|
||
|
identity string
|
||
|
stopC chan struct{}
|
||
|
}
|
||
|
|
||
|
// ActivityWorker wraps the code for hosting activity types.
|
||
|
// TODO: Worker doing heartbeating automatically while activity task is running
|
||
|
activityWorker struct {
|
||
|
executionParameters workerExecutionParameters
|
||
|
workflowService workflowservice.WorkflowServiceClient
|
||
|
poller taskPoller
|
||
|
worker *baseWorker
|
||
|
identity string
|
||
|
stopC chan struct{}
|
||
|
}
|
||
|
|
||
|
// sessionWorker wraps the code for hosting session creation, completion and
|
||
|
// activities within a session. The creationWorker polls from a global taskqueue,
|
||
|
// while the activityWorker polls from a resource specific taskqueue.
|
||
|
sessionWorker struct {
|
||
|
creationWorker *activityWorker
|
||
|
activityWorker *activityWorker
|
||
|
}
|
||
|
|
||
|
// Worker overrides.
|
||
|
workerOverrides struct {
|
||
|
workflowTaskHandler WorkflowTaskHandler
|
||
|
activityTaskHandler ActivityTaskHandler
|
||
|
}
|
||
|
|
||
|
// workerExecutionParameters defines worker configure/execution options.
|
||
|
workerExecutionParameters struct {
|
||
|
// Namespace name.
|
||
|
Namespace string
|
||
|
|
||
|
// Task queue name to poll.
|
||
|
TaskQueue string
|
||
|
|
||
|
// Defines how many concurrent activity executions by this worker.
|
||
|
ConcurrentActivityExecutionSize int
|
||
|
|
||
|
// Defines rate limiting on number of activity tasks that can be executed per second per worker.
|
||
|
WorkerActivitiesPerSecond float64
|
||
|
|
||
|
// MaxConcurrentActivityTaskQueuePollers is the max number of pollers for activity task queue.
|
||
|
MaxConcurrentActivityTaskQueuePollers int
|
||
|
|
||
|
// Defines how many concurrent workflow task executions by this worker.
|
||
|
ConcurrentWorkflowTaskExecutionSize int
|
||
|
|
||
|
// MaxConcurrentWorkflowTaskQueuePollers is the max number of pollers for workflow task queue.
|
||
|
MaxConcurrentWorkflowTaskQueuePollers int
|
||
|
|
||
|
// Defines how many concurrent local activity executions by this worker.
|
||
|
ConcurrentLocalActivityExecutionSize int
|
||
|
|
||
|
// Defines rate limiting on number of local activities that can be executed per second per worker.
|
||
|
WorkerLocalActivitiesPerSecond float64
|
||
|
|
||
|
// TaskQueueActivitiesPerSecond is the throttling limit for activity tasks controlled by the server.
|
||
|
TaskQueueActivitiesPerSecond float64
|
||
|
|
||
|
// User can provide an identity for the debuggability. If not provided the framework has
|
||
|
// a default option.
|
||
|
Identity string
|
||
|
|
||
|
MetricsHandler metrics.Handler
|
||
|
|
||
|
Logger log.Logger
|
||
|
|
||
|
// Enable logging in replay mode
|
||
|
EnableLoggingInReplay bool
|
||
|
|
||
|
// Context to store user provided key/value pairs
|
||
|
UserContext context.Context
|
||
|
|
||
|
// Context cancel function to cancel user context
|
||
|
UserContextCancel context.CancelFunc
|
||
|
|
||
|
StickyScheduleToStartTimeout time.Duration
|
||
|
|
||
|
// WorkflowPanicPolicy is used for configuring how client's workflow task handler deals with workflow
|
||
|
// code panicking which includes non backwards compatible changes to the workflow code without appropriate
|
||
|
// versioning (see workflow.GetVersion).
|
||
|
// The default behavior is to block workflow execution until the problem is fixed.
|
||
|
WorkflowPanicPolicy WorkflowPanicPolicy
|
||
|
|
||
|
DataConverter converter.DataConverter
|
||
|
|
||
|
// WorkerStopTimeout is the time delay before hard terminate worker
|
||
|
WorkerStopTimeout time.Duration
|
||
|
|
||
|
// WorkerStopChannel is a read only channel listen on worker close. The worker will close the channel before exit.
|
||
|
WorkerStopChannel <-chan struct{}
|
||
|
|
||
|
// SessionResourceID is a unique identifier of the resource the session will consume
|
||
|
SessionResourceID string
|
||
|
|
||
|
ContextPropagators []ContextPropagator
|
||
|
|
||
|
// DeadlockDetectionTimeout specifies workflow task timeout.
|
||
|
DeadlockDetectionTimeout time.Duration
|
||
|
|
||
|
DefaultHeartbeatThrottleInterval time.Duration
|
||
|
|
||
|
MaxHeartbeatThrottleInterval time.Duration
|
||
|
|
||
|
// Pointer to the shared worker cache
|
||
|
cache *WorkerCache
|
||
|
}
|
||
|
)
|
||
|
|
||
|
var debugMode = os.Getenv("TEMPORAL_DEBUG") != ""
|
||
|
|
||
|
// newWorkflowWorker returns an instance of the workflow worker.
|
||
|
func newWorkflowWorker(service workflowservice.WorkflowServiceClient, params workerExecutionParameters, ppMgr pressurePointMgr, registry *registry) *workflowWorker {
|
||
|
return newWorkflowWorkerInternal(service, params, ppMgr, nil, registry)
|
||
|
}
|
||
|
|
||
|
func ensureRequiredParams(params *workerExecutionParameters) {
|
||
|
if params.Identity == "" {
|
||
|
params.Identity = getWorkerIdentity(params.TaskQueue)
|
||
|
}
|
||
|
if params.Logger == nil {
|
||
|
// create default logger if user does not supply one (should happen in tests only).
|
||
|
params.Logger = ilog.NewDefaultLogger()
|
||
|
params.Logger.Info("No logger configured for temporal worker. Created default one.")
|
||
|
}
|
||
|
if params.MetricsHandler == nil {
|
||
|
params.MetricsHandler = metrics.NopHandler
|
||
|
params.Logger.Info("No metrics handler configured for temporal worker. Use NopHandler as default.")
|
||
|
}
|
||
|
if params.DataConverter == nil {
|
||
|
params.DataConverter = converter.GetDefaultDataConverter()
|
||
|
params.Logger.Info("No DataConverter configured for temporal worker. Use default one.")
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// verifyNamespaceExist does a DescribeNamespace operation on the specified namespace with backoff/retry
|
||
|
func verifyNamespaceExist(
|
||
|
client workflowservice.WorkflowServiceClient,
|
||
|
metricsHandler metrics.Handler,
|
||
|
namespace string,
|
||
|
logger log.Logger,
|
||
|
) error {
|
||
|
ctx := context.Background()
|
||
|
if namespace == "" {
|
||
|
return errors.New("namespace cannot be empty")
|
||
|
}
|
||
|
grpcCtx, cancel := newGRPCContext(ctx, grpcMetricsHandler(metricsHandler), defaultGrpcRetryParameters(ctx))
|
||
|
defer cancel()
|
||
|
_, err := client.DescribeNamespace(grpcCtx, &workflowservice.DescribeNamespaceRequest{Namespace: namespace})
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
func newWorkflowWorkerInternal(service workflowservice.WorkflowServiceClient, params workerExecutionParameters, ppMgr pressurePointMgr, overrides *workerOverrides, registry *registry) *workflowWorker {
|
||
|
workerStopChannel := make(chan struct{})
|
||
|
params.WorkerStopChannel = getReadOnlyChannel(workerStopChannel)
|
||
|
// Get a workflow task handler.
|
||
|
ensureRequiredParams(¶ms)
|
||
|
var taskHandler WorkflowTaskHandler
|
||
|
if overrides != nil && overrides.workflowTaskHandler != nil {
|
||
|
taskHandler = overrides.workflowTaskHandler
|
||
|
} else {
|
||
|
taskHandler = newWorkflowTaskHandler(params, ppMgr, registry)
|
||
|
}
|
||
|
return newWorkflowTaskWorkerInternal(taskHandler, service, params, workerStopChannel, registry.interceptors)
|
||
|
}
|
||
|
|
||
|
func newWorkflowTaskWorkerInternal(
|
||
|
taskHandler WorkflowTaskHandler,
|
||
|
service workflowservice.WorkflowServiceClient,
|
||
|
params workerExecutionParameters,
|
||
|
stopC chan struct{},
|
||
|
interceptors []WorkerInterceptor,
|
||
|
) *workflowWorker {
|
||
|
ensureRequiredParams(¶ms)
|
||
|
poller := newWorkflowTaskPoller(taskHandler, service, params)
|
||
|
worker := newBaseWorker(baseWorkerOptions{
|
||
|
pollerCount: params.MaxConcurrentWorkflowTaskQueuePollers,
|
||
|
pollerRate: defaultPollerRate,
|
||
|
maxConcurrentTask: params.ConcurrentWorkflowTaskExecutionSize,
|
||
|
maxTaskPerSecond: defaultWorkerTaskExecutionRate,
|
||
|
taskWorker: poller,
|
||
|
identity: params.Identity,
|
||
|
workerType: "WorkflowWorker",
|
||
|
stopTimeout: params.WorkerStopTimeout},
|
||
|
params.Logger,
|
||
|
params.MetricsHandler,
|
||
|
nil,
|
||
|
)
|
||
|
|
||
|
// laTunnel is the glue that hookup 3 parts
|
||
|
laTunnel := newLocalActivityTunnel(params.WorkerStopChannel)
|
||
|
|
||
|
// 1) workflow handler will send local activity task to laTunnel
|
||
|
if handlerImpl, ok := taskHandler.(*workflowTaskHandlerImpl); ok {
|
||
|
handlerImpl.laTunnel = laTunnel
|
||
|
}
|
||
|
|
||
|
// 2) local activity task poller will poll from laTunnel, and result will be pushed to laTunnel
|
||
|
localActivityTaskPoller := newLocalActivityPoller(params, laTunnel, interceptors)
|
||
|
localActivityWorker := newBaseWorker(baseWorkerOptions{
|
||
|
pollerCount: 1, // 1 poller (from local channel) is enough for local activity
|
||
|
maxConcurrentTask: params.ConcurrentLocalActivityExecutionSize,
|
||
|
maxTaskPerSecond: params.WorkerLocalActivitiesPerSecond,
|
||
|
taskWorker: localActivityTaskPoller,
|
||
|
identity: params.Identity,
|
||
|
workerType: "LocalActivityWorker",
|
||
|
stopTimeout: params.WorkerStopTimeout},
|
||
|
params.Logger,
|
||
|
params.MetricsHandler,
|
||
|
nil,
|
||
|
)
|
||
|
|
||
|
// 3) the result pushed to laTunnel will be send as task to workflow worker to process.
|
||
|
worker.taskQueueCh = laTunnel.resultCh
|
||
|
|
||
|
return &workflowWorker{
|
||
|
executionParameters: params,
|
||
|
workflowService: service,
|
||
|
poller: poller,
|
||
|
worker: worker,
|
||
|
localActivityWorker: localActivityWorker,
|
||
|
identity: params.Identity,
|
||
|
stopC: stopC,
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Start the worker.
|
||
|
func (ww *workflowWorker) Start() error {
|
||
|
err := verifyNamespaceExist(ww.workflowService, ww.executionParameters.MetricsHandler, ww.executionParameters.Namespace, ww.worker.logger)
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
ww.localActivityWorker.Start()
|
||
|
ww.worker.Start()
|
||
|
return nil // TODO: propagate error
|
||
|
}
|
||
|
|
||
|
// Stop the worker.
|
||
|
func (ww *workflowWorker) Stop() {
|
||
|
close(ww.stopC)
|
||
|
// TODO: remove the stop methods in favor of the workerStopChannel
|
||
|
ww.localActivityWorker.Stop()
|
||
|
ww.worker.Stop()
|
||
|
}
|
||
|
|
||
|
func newSessionWorker(service workflowservice.WorkflowServiceClient, params workerExecutionParameters, overrides *workerOverrides, env *registry, maxConcurrentSessionExecutionSize int) *sessionWorker {
|
||
|
if params.Identity == "" {
|
||
|
params.Identity = getWorkerIdentity(params.TaskQueue)
|
||
|
}
|
||
|
// For now resourceID is hidden from user so we will always create a unique one for each worker.
|
||
|
if params.SessionResourceID == "" {
|
||
|
params.SessionResourceID = uuid.New()
|
||
|
}
|
||
|
sessionEnvironment := newSessionEnvironment(params.SessionResourceID, maxConcurrentSessionExecutionSize)
|
||
|
|
||
|
creationTaskqueue := getCreationTaskqueue(params.TaskQueue)
|
||
|
params.UserContext = context.WithValue(params.UserContext, sessionEnvironmentContextKey, sessionEnvironment)
|
||
|
params.TaskQueue = sessionEnvironment.GetResourceSpecificTaskqueue()
|
||
|
activityWorker := newActivityWorker(service, params, overrides, env, nil)
|
||
|
|
||
|
params.MaxConcurrentActivityTaskQueuePollers = 1
|
||
|
params.TaskQueue = creationTaskqueue
|
||
|
creationWorker := newActivityWorker(service, params, overrides, env, sessionEnvironment.GetTokenBucket())
|
||
|
|
||
|
return &sessionWorker{
|
||
|
creationWorker: creationWorker,
|
||
|
activityWorker: activityWorker,
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func (sw *sessionWorker) Start() error {
|
||
|
err := sw.creationWorker.Start()
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
err = sw.activityWorker.Start()
|
||
|
if err != nil {
|
||
|
sw.creationWorker.Stop()
|
||
|
return err
|
||
|
}
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
func (sw *sessionWorker) Stop() {
|
||
|
sw.creationWorker.Stop()
|
||
|
sw.activityWorker.Stop()
|
||
|
}
|
||
|
|
||
|
func newActivityWorker(service workflowservice.WorkflowServiceClient, params workerExecutionParameters, overrides *workerOverrides, env *registry, sessionTokenBucket *sessionTokenBucket) *activityWorker {
|
||
|
workerStopChannel := make(chan struct{}, 1)
|
||
|
params.WorkerStopChannel = getReadOnlyChannel(workerStopChannel)
|
||
|
ensureRequiredParams(¶ms)
|
||
|
|
||
|
// Get a activity task handler.
|
||
|
var taskHandler ActivityTaskHandler
|
||
|
if overrides != nil && overrides.activityTaskHandler != nil {
|
||
|
taskHandler = overrides.activityTaskHandler
|
||
|
} else {
|
||
|
taskHandler = newActivityTaskHandler(service, params, env)
|
||
|
}
|
||
|
return newActivityTaskWorker(taskHandler, service, params, sessionTokenBucket, workerStopChannel)
|
||
|
}
|
||
|
|
||
|
func newActivityTaskWorker(taskHandler ActivityTaskHandler, service workflowservice.WorkflowServiceClient, workerParams workerExecutionParameters, sessionTokenBucket *sessionTokenBucket, stopC chan struct{}) (worker *activityWorker) {
|
||
|
ensureRequiredParams(&workerParams)
|
||
|
|
||
|
poller := newActivityTaskPoller(taskHandler, service, workerParams)
|
||
|
|
||
|
base := newBaseWorker(
|
||
|
baseWorkerOptions{
|
||
|
pollerCount: workerParams.MaxConcurrentActivityTaskQueuePollers,
|
||
|
pollerRate: defaultPollerRate,
|
||
|
maxConcurrentTask: workerParams.ConcurrentActivityExecutionSize,
|
||
|
maxTaskPerSecond: workerParams.WorkerActivitiesPerSecond,
|
||
|
taskWorker: poller,
|
||
|
identity: workerParams.Identity,
|
||
|
workerType: "ActivityWorker",
|
||
|
stopTimeout: workerParams.WorkerStopTimeout,
|
||
|
userContextCancel: workerParams.UserContextCancel},
|
||
|
workerParams.Logger,
|
||
|
workerParams.MetricsHandler,
|
||
|
sessionTokenBucket,
|
||
|
)
|
||
|
|
||
|
return &activityWorker{
|
||
|
executionParameters: workerParams,
|
||
|
workflowService: service,
|
||
|
worker: base,
|
||
|
poller: poller,
|
||
|
identity: workerParams.Identity,
|
||
|
stopC: stopC,
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Start the worker.
|
||
|
func (aw *activityWorker) Start() error {
|
||
|
err := verifyNamespaceExist(aw.workflowService, aw.executionParameters.MetricsHandler, aw.executionParameters.Namespace, aw.worker.logger)
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
aw.worker.Start()
|
||
|
return nil // TODO: propagate errors
|
||
|
}
|
||
|
|
||
|
// Stop the worker.
|
||
|
func (aw *activityWorker) Stop() {
|
||
|
close(aw.stopC)
|
||
|
aw.worker.Stop()
|
||
|
}
|
||
|
|
||
|
type registry struct {
|
||
|
sync.Mutex
|
||
|
workflowFuncMap map[string]interface{}
|
||
|
workflowAliasMap map[string]string
|
||
|
activityFuncMap map[string]activity
|
||
|
activityAliasMap map[string]string
|
||
|
interceptors []WorkerInterceptor
|
||
|
}
|
||
|
|
||
|
func (r *registry) RegisterWorkflow(af interface{}) {
|
||
|
r.RegisterWorkflowWithOptions(af, RegisterWorkflowOptions{})
|
||
|
}
|
||
|
|
||
|
func (r *registry) RegisterWorkflowWithOptions(
|
||
|
wf interface{},
|
||
|
options RegisterWorkflowOptions,
|
||
|
) {
|
||
|
// Support direct registration of WorkflowDefinition
|
||
|
factory, ok := wf.(WorkflowDefinitionFactory)
|
||
|
if ok {
|
||
|
if len(options.Name) == 0 {
|
||
|
panic("WorkflowDefinitionFactory must be registered with a name")
|
||
|
}
|
||
|
r.workflowFuncMap[options.Name] = factory
|
||
|
return
|
||
|
}
|
||
|
// Validate that it is a function
|
||
|
fnType := reflect.TypeOf(wf)
|
||
|
if err := validateFnFormat(fnType, true); err != nil {
|
||
|
panic(err)
|
||
|
}
|
||
|
fnName, _ := getFunctionName(wf)
|
||
|
alias := options.Name
|
||
|
registerName := fnName
|
||
|
if len(alias) > 0 {
|
||
|
registerName = alias
|
||
|
}
|
||
|
|
||
|
r.Lock()
|
||
|
defer r.Unlock()
|
||
|
|
||
|
if !options.DisableAlreadyRegisteredCheck {
|
||
|
if _, ok := r.workflowFuncMap[registerName]; ok {
|
||
|
panic(fmt.Sprintf("workflow name \"%v\" is already registered", registerName))
|
||
|
}
|
||
|
}
|
||
|
r.workflowFuncMap[registerName] = wf
|
||
|
if len(alias) > 0 {
|
||
|
r.workflowAliasMap[fnName] = alias
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func (r *registry) RegisterActivity(af interface{}) {
|
||
|
r.RegisterActivityWithOptions(af, RegisterActivityOptions{})
|
||
|
}
|
||
|
|
||
|
func (r *registry) RegisterActivityWithOptions(
|
||
|
af interface{},
|
||
|
options RegisterActivityOptions,
|
||
|
) {
|
||
|
// Support direct registration of activity
|
||
|
a, ok := af.(activity)
|
||
|
if ok {
|
||
|
if options.Name == "" {
|
||
|
panic("registration of activity interface requires name")
|
||
|
}
|
||
|
r.addActivityWithLock(options.Name, a)
|
||
|
return
|
||
|
}
|
||
|
// Validate that it is a function
|
||
|
fnType := reflect.TypeOf(af)
|
||
|
if fnType.Kind() == reflect.Ptr && fnType.Elem().Kind() == reflect.Struct {
|
||
|
registerErr := r.registerActivityStructWithOptions(af, options)
|
||
|
if registerErr != nil {
|
||
|
panic(registerErr)
|
||
|
}
|
||
|
return
|
||
|
}
|
||
|
if err := validateFnFormat(fnType, false); err != nil {
|
||
|
panic(err)
|
||
|
}
|
||
|
fnName, _ := getFunctionName(af)
|
||
|
alias := options.Name
|
||
|
registerName := fnName
|
||
|
if len(alias) > 0 {
|
||
|
registerName = alias
|
||
|
}
|
||
|
|
||
|
r.Lock()
|
||
|
defer r.Unlock()
|
||
|
|
||
|
if !options.DisableAlreadyRegisteredCheck {
|
||
|
if _, ok := r.activityFuncMap[registerName]; ok {
|
||
|
panic(fmt.Sprintf("activity type \"%v\" is already registered", registerName))
|
||
|
}
|
||
|
}
|
||
|
r.activityFuncMap[registerName] = &activityExecutor{name: registerName, fn: af}
|
||
|
if len(alias) > 0 {
|
||
|
r.activityAliasMap[fnName] = alias
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func (r *registry) registerActivityStructWithOptions(aStruct interface{}, options RegisterActivityOptions) error {
|
||
|
r.Lock()
|
||
|
defer r.Unlock()
|
||
|
|
||
|
structValue := reflect.ValueOf(aStruct)
|
||
|
structType := structValue.Type()
|
||
|
count := 0
|
||
|
for i := 0; i < structValue.NumMethod(); i++ {
|
||
|
methodValue := structValue.Method(i)
|
||
|
method := structType.Method(i)
|
||
|
// skip private method
|
||
|
if method.PkgPath != "" {
|
||
|
continue
|
||
|
}
|
||
|
name := method.Name
|
||
|
if err := validateFnFormat(method.Type, false); err != nil {
|
||
|
if options.SkipInvalidStructFunctions {
|
||
|
continue
|
||
|
}
|
||
|
|
||
|
return fmt.Errorf("method %s of %s: %w", name, structType.Name(), err)
|
||
|
}
|
||
|
registerName := options.Name + name
|
||
|
if !options.DisableAlreadyRegisteredCheck {
|
||
|
if _, ok := r.getActivityNoLock(registerName); ok {
|
||
|
return fmt.Errorf("activity type \"%v\" is already registered", registerName)
|
||
|
}
|
||
|
}
|
||
|
r.activityFuncMap[registerName] = &activityExecutor{name: registerName, fn: methodValue.Interface()}
|
||
|
count++
|
||
|
}
|
||
|
if count == 0 {
|
||
|
return fmt.Errorf("no activities (public methods) found at %v structure", structType.Name())
|
||
|
}
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
func (r *registry) getWorkflowAlias(fnName string) (string, bool) {
|
||
|
r.Lock()
|
||
|
defer r.Unlock()
|
||
|
alias, ok := r.workflowAliasMap[fnName]
|
||
|
return alias, ok
|
||
|
}
|
||
|
|
||
|
func (r *registry) getWorkflowFn(fnName string) (interface{}, bool) {
|
||
|
r.Lock()
|
||
|
defer r.Unlock()
|
||
|
fn, ok := r.workflowFuncMap[fnName]
|
||
|
return fn, ok
|
||
|
}
|
||
|
|
||
|
func (r *registry) getRegisteredWorkflowTypes() []string {
|
||
|
r.Lock()
|
||
|
defer r.Unlock()
|
||
|
var result []string
|
||
|
for t := range r.workflowFuncMap {
|
||
|
result = append(result, t)
|
||
|
}
|
||
|
return result
|
||
|
}
|
||
|
|
||
|
func (r *registry) getActivityAlias(fnName string) (string, bool) {
|
||
|
r.Lock()
|
||
|
defer r.Unlock()
|
||
|
alias, ok := r.activityAliasMap[fnName]
|
||
|
return alias, ok
|
||
|
}
|
||
|
|
||
|
func (r *registry) addActivityWithLock(fnName string, a activity) {
|
||
|
r.Lock()
|
||
|
defer r.Unlock()
|
||
|
r.activityFuncMap[fnName] = a
|
||
|
}
|
||
|
|
||
|
func (r *registry) GetActivity(fnName string) (activity, bool) {
|
||
|
r.Lock()
|
||
|
defer r.Unlock()
|
||
|
a, ok := r.activityFuncMap[fnName]
|
||
|
return a, ok
|
||
|
}
|
||
|
|
||
|
func (r *registry) getActivityNoLock(fnName string) (activity, bool) {
|
||
|
a, ok := r.activityFuncMap[fnName]
|
||
|
return a, ok
|
||
|
}
|
||
|
|
||
|
func (r *registry) getRegisteredActivities() []activity {
|
||
|
r.Lock()
|
||
|
defer r.Unlock()
|
||
|
activities := make([]activity, 0, len(r.activityFuncMap))
|
||
|
for _, a := range r.activityFuncMap {
|
||
|
activities = append(activities, a)
|
||
|
}
|
||
|
return activities
|
||
|
}
|
||
|
|
||
|
func (r *registry) getRegisteredActivityTypes() []string {
|
||
|
r.Lock()
|
||
|
defer r.Unlock()
|
||
|
var result []string
|
||
|
for name := range r.activityFuncMap {
|
||
|
result = append(result, name)
|
||
|
}
|
||
|
return result
|
||
|
}
|
||
|
|
||
|
func (r *registry) getWorkflowDefinition(wt WorkflowType) (WorkflowDefinition, error) {
|
||
|
lookup := wt.Name
|
||
|
if alias, ok := r.getWorkflowAlias(lookup); ok {
|
||
|
lookup = alias
|
||
|
}
|
||
|
wf, ok := r.getWorkflowFn(lookup)
|
||
|
if !ok {
|
||
|
supported := strings.Join(r.getRegisteredWorkflowTypes(), ", ")
|
||
|
return nil, fmt.Errorf("unable to find workflow type: %v. Supported types: [%v]", lookup, supported)
|
||
|
}
|
||
|
wdf, ok := wf.(WorkflowDefinitionFactory)
|
||
|
if ok {
|
||
|
return wdf.NewWorkflowDefinition(), nil
|
||
|
}
|
||
|
executor := &workflowExecutor{workflowType: lookup, fn: wf, interceptors: r.interceptors}
|
||
|
return newSyncWorkflowDefinition(executor), nil
|
||
|
}
|
||
|
|
||
|
// Validate function parameters.
|
||
|
func validateFnFormat(fnType reflect.Type, isWorkflow bool) error {
|
||
|
if fnType.Kind() != reflect.Func {
|
||
|
return fmt.Errorf("expected a func as input but was %s", fnType.Kind())
|
||
|
}
|
||
|
if isWorkflow {
|
||
|
if fnType.NumIn() < 1 {
|
||
|
return fmt.Errorf(
|
||
|
"expected at least one argument of type workflow.Context in function, found %d input arguments",
|
||
|
fnType.NumIn(),
|
||
|
)
|
||
|
}
|
||
|
if !isWorkflowContext(fnType.In(0)) {
|
||
|
return fmt.Errorf("expected first argument to be workflow.Context but found %s", fnType.In(0))
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Return values
|
||
|
// We expect either
|
||
|
// <result>, error
|
||
|
// (or) just error
|
||
|
if fnType.NumOut() < 1 || fnType.NumOut() > 2 {
|
||
|
return fmt.Errorf(
|
||
|
"expected function to return result, error or just error, but found %d return values", fnType.NumOut(),
|
||
|
)
|
||
|
}
|
||
|
if fnType.NumOut() > 1 && !isValidResultType(fnType.Out(0)) {
|
||
|
return fmt.Errorf(
|
||
|
"expected function first return value to return valid type but found: %v", fnType.Out(0).Kind(),
|
||
|
)
|
||
|
}
|
||
|
if !isError(fnType.Out(fnType.NumOut() - 1)) {
|
||
|
return fmt.Errorf(
|
||
|
"expected function second return value to return error but found %v", fnType.Out(fnType.NumOut()-1).Kind(),
|
||
|
)
|
||
|
}
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
func newRegistry() *registry {
|
||
|
return ®istry{
|
||
|
workflowFuncMap: make(map[string]interface{}),
|
||
|
workflowAliasMap: make(map[string]string),
|
||
|
activityFuncMap: make(map[string]activity),
|
||
|
activityAliasMap: make(map[string]string),
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Wrapper to execute workflow functions.
|
||
|
type workflowExecutor struct {
|
||
|
workflowType string
|
||
|
fn interface{}
|
||
|
interceptors []WorkerInterceptor
|
||
|
}
|
||
|
|
||
|
func (we *workflowExecutor) Execute(ctx Context, input *commonpb.Payloads) (*commonpb.Payloads, error) {
|
||
|
dataConverter := WithWorkflowContext(ctx, getWorkflowEnvOptions(ctx).DataConverter)
|
||
|
fnType := reflect.TypeOf(we.fn)
|
||
|
|
||
|
args, err := decodeArgsToRawValues(dataConverter, fnType, input)
|
||
|
if err != nil {
|
||
|
return nil, fmt.Errorf(
|
||
|
"unable to decode the workflow function input payload with error: %w, function name: %v",
|
||
|
err, we.workflowType)
|
||
|
}
|
||
|
|
||
|
envInterceptor := getWorkflowEnvironmentInterceptor(ctx)
|
||
|
envInterceptor.fn = we.fn
|
||
|
|
||
|
// Execute and serialize result
|
||
|
result, err := envInterceptor.inboundInterceptor.ExecuteWorkflow(ctx, &ExecuteWorkflowInput{Args: args})
|
||
|
var serializedResult *commonpb.Payloads
|
||
|
if err == nil && result != nil {
|
||
|
serializedResult, err = encodeArg(dataConverter, result)
|
||
|
}
|
||
|
return serializedResult, err
|
||
|
}
|
||
|
|
||
|
// Wrapper to execute activity functions.
|
||
|
type activityExecutor struct {
|
||
|
name string
|
||
|
fn interface{}
|
||
|
skipInterceptors bool
|
||
|
}
|
||
|
|
||
|
func (ae *activityExecutor) ActivityType() ActivityType {
|
||
|
return ActivityType{Name: ae.name}
|
||
|
}
|
||
|
|
||
|
func (ae *activityExecutor) GetFunction() interface{} {
|
||
|
return ae.fn
|
||
|
}
|
||
|
|
||
|
func (ae *activityExecutor) Execute(ctx context.Context, input *commonpb.Payloads) (*commonpb.Payloads, error) {
|
||
|
fnType := reflect.TypeOf(ae.fn)
|
||
|
dataConverter := getDataConverterFromActivityCtx(ctx)
|
||
|
|
||
|
args, err := decodeArgsToRawValues(dataConverter, fnType, input)
|
||
|
if err != nil {
|
||
|
return nil, fmt.Errorf(
|
||
|
"unable to decode the activity function input payload with error: %w for function name: %v",
|
||
|
err, ae.name)
|
||
|
}
|
||
|
|
||
|
return ae.ExecuteWithActualArgs(ctx, args)
|
||
|
}
|
||
|
|
||
|
func (ae *activityExecutor) ExecuteWithActualArgs(ctx context.Context, args []interface{}) (*commonpb.Payloads, error) {
|
||
|
dataConverter := getDataConverterFromActivityCtx(ctx)
|
||
|
|
||
|
envInterceptor := getActivityEnvironmentInterceptor(ctx)
|
||
|
envInterceptor.fn = ae.fn
|
||
|
|
||
|
// Execute and serialize result
|
||
|
interceptor := envInterceptor.inboundInterceptor
|
||
|
if ae.skipInterceptors {
|
||
|
interceptor = envInterceptor
|
||
|
}
|
||
|
result, resultErr := interceptor.ExecuteActivity(ctx, &ExecuteActivityInput{Args: args})
|
||
|
var serializedResult *commonpb.Payloads
|
||
|
if result != nil {
|
||
|
// As a special case, if the result is already a payload, just use it
|
||
|
var ok bool
|
||
|
if serializedResult, ok = result.(*commonpb.Payloads); !ok {
|
||
|
var err error
|
||
|
if serializedResult, err = encodeArg(dataConverter, result); err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
return serializedResult, resultErr
|
||
|
}
|
||
|
|
||
|
func getDataConverterFromActivityCtx(ctx context.Context) converter.DataConverter {
|
||
|
var dataConverter converter.DataConverter
|
||
|
|
||
|
env := getActivityEnvironmentFromCtx(ctx)
|
||
|
if env != nil && env.dataConverter != nil {
|
||
|
dataConverter = env.dataConverter
|
||
|
} else {
|
||
|
dataConverter = converter.GetDefaultDataConverter()
|
||
|
}
|
||
|
return WithContext(ctx, dataConverter)
|
||
|
}
|
||
|
|
||
|
func getNamespaceFromActivityCtx(ctx context.Context) string {
|
||
|
env := getActivityEnvironmentFromCtx(ctx)
|
||
|
if env == nil {
|
||
|
return ""
|
||
|
}
|
||
|
return env.workflowNamespace
|
||
|
}
|
||
|
|
||
|
func getActivityEnvironmentFromCtx(ctx context.Context) *activityEnvironment {
|
||
|
if ctx == nil || ctx.Value(activityEnvContextKey) == nil {
|
||
|
return nil
|
||
|
}
|
||
|
return ctx.Value(activityEnvContextKey).(*activityEnvironment)
|
||
|
}
|
||
|
|
||
|
// AggregatedWorker combines management of both workflowWorker and activityWorker worker lifecycle.
|
||
|
type AggregatedWorker struct {
|
||
|
workflowWorker *workflowWorker
|
||
|
activityWorker *activityWorker
|
||
|
sessionWorker *sessionWorker
|
||
|
logger log.Logger
|
||
|
registry *registry
|
||
|
stopC chan struct{}
|
||
|
}
|
||
|
|
||
|
// RegisterWorkflow registers workflow implementation with the AggregatedWorker
|
||
|
func (aw *AggregatedWorker) RegisterWorkflow(w interface{}) {
|
||
|
if aw.workflowWorker == nil {
|
||
|
panic("workflow worker disabled, cannot register workflow")
|
||
|
}
|
||
|
aw.registry.RegisterWorkflow(w)
|
||
|
}
|
||
|
|
||
|
// RegisterWorkflowWithOptions registers workflow implementation with the AggregatedWorker
|
||
|
func (aw *AggregatedWorker) RegisterWorkflowWithOptions(w interface{}, options RegisterWorkflowOptions) {
|
||
|
if aw.workflowWorker == nil {
|
||
|
panic("workflow worker disabled, cannot register workflow")
|
||
|
}
|
||
|
aw.registry.RegisterWorkflowWithOptions(w, options)
|
||
|
}
|
||
|
|
||
|
// RegisterActivity registers activity implementation with the AggregatedWorker
|
||
|
func (aw *AggregatedWorker) RegisterActivity(a interface{}) {
|
||
|
aw.registry.RegisterActivity(a)
|
||
|
}
|
||
|
|
||
|
// RegisterActivityWithOptions registers activity implementation with the AggregatedWorker
|
||
|
func (aw *AggregatedWorker) RegisterActivityWithOptions(a interface{}, options RegisterActivityOptions) {
|
||
|
aw.registry.RegisterActivityWithOptions(a, options)
|
||
|
}
|
||
|
|
||
|
// Start the worker in a non-blocking fashion.
|
||
|
func (aw *AggregatedWorker) Start() error {
|
||
|
aw.assertNotStopped()
|
||
|
if err := initBinaryChecksum(); err != nil {
|
||
|
return fmt.Errorf("failed to get executable checksum: %v", err)
|
||
|
}
|
||
|
|
||
|
if !util.IsInterfaceNil(aw.workflowWorker) {
|
||
|
if err := aw.workflowWorker.Start(); err != nil {
|
||
|
return err
|
||
|
}
|
||
|
}
|
||
|
if !util.IsInterfaceNil(aw.activityWorker) {
|
||
|
if err := aw.activityWorker.Start(); err != nil {
|
||
|
// stop workflow worker.
|
||
|
if !util.IsInterfaceNil(aw.workflowWorker) {
|
||
|
if aw.workflowWorker.worker.isWorkerStarted {
|
||
|
aw.workflowWorker.Stop()
|
||
|
}
|
||
|
}
|
||
|
return err
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if !util.IsInterfaceNil(aw.sessionWorker) && len(aw.registry.getRegisteredActivities()) > 0 {
|
||
|
aw.logger.Info("Starting session worker")
|
||
|
if err := aw.sessionWorker.Start(); err != nil {
|
||
|
// stop workflow worker and activity worker.
|
||
|
if !util.IsInterfaceNil(aw.workflowWorker) {
|
||
|
if aw.workflowWorker.worker.isWorkerStarted {
|
||
|
aw.workflowWorker.Stop()
|
||
|
}
|
||
|
}
|
||
|
if !util.IsInterfaceNil(aw.activityWorker) {
|
||
|
if aw.activityWorker.worker.isWorkerStarted {
|
||
|
aw.activityWorker.Stop()
|
||
|
}
|
||
|
}
|
||
|
return err
|
||
|
}
|
||
|
}
|
||
|
aw.logger.Info("Started Worker")
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
func (aw *AggregatedWorker) assertNotStopped() {
|
||
|
stopped := true
|
||
|
select {
|
||
|
case <-aw.stopC:
|
||
|
default:
|
||
|
stopped = false
|
||
|
}
|
||
|
if stopped {
|
||
|
panic("attempted to start a worker that has been stopped before")
|
||
|
}
|
||
|
}
|
||
|
|
||
|
var binaryChecksum string
|
||
|
var binaryChecksumLock sync.Mutex
|
||
|
|
||
|
// SetBinaryChecksum sets the identifier of the binary(aka BinaryChecksum).
|
||
|
// The identifier is mainly used in recording reset points when respondWorkflowTaskCompleted. For each workflow, the very first
|
||
|
// workflow task completed by a binary will be associated as a auto-reset point for the binary. So that when a customer wants to
|
||
|
// mark the binary as bad, the workflow will be reset to that point -- which means workflow will forget all progress generated
|
||
|
// by the binary.
|
||
|
// On another hand, once the binary is marked as bad, the bad binary cannot poll workflow queue and make any progress any more.
|
||
|
func SetBinaryChecksum(checksum string) {
|
||
|
binaryChecksumLock.Lock()
|
||
|
defer binaryChecksumLock.Unlock()
|
||
|
|
||
|
binaryChecksum = checksum
|
||
|
}
|
||
|
|
||
|
func initBinaryChecksum() error {
|
||
|
binaryChecksumLock.Lock()
|
||
|
defer binaryChecksumLock.Unlock()
|
||
|
|
||
|
return initBinaryChecksumLocked()
|
||
|
}
|
||
|
|
||
|
// callers MUST hold binaryChecksumLock before calling
|
||
|
func initBinaryChecksumLocked() error {
|
||
|
if len(binaryChecksum) > 0 {
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
exec, err := os.Executable()
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
f, err := os.Open(exec)
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
defer func() {
|
||
|
_ = f.Close() // error is unimportant as it is read-only
|
||
|
}()
|
||
|
|
||
|
h := md5.New()
|
||
|
if _, err := io.Copy(h, f); err != nil {
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
checksum := h.Sum(nil)
|
||
|
binaryChecksum = hex.EncodeToString(checksum[:])
|
||
|
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
func getBinaryChecksum() string {
|
||
|
binaryChecksumLock.Lock()
|
||
|
defer binaryChecksumLock.Unlock()
|
||
|
|
||
|
if len(binaryChecksum) == 0 {
|
||
|
err := initBinaryChecksumLocked()
|
||
|
if err != nil {
|
||
|
panic(err)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
return binaryChecksum
|
||
|
}
|
||
|
|
||
|
// Run the worker in a blocking fashion. Stop the worker when interruptCh receives signal.
|
||
|
// Pass worker.InterruptCh() to stop the worker with SIGINT or SIGTERM.
|
||
|
// Pass nil to stop the worker with external Stop() call.
|
||
|
// Pass any other `<-chan interface{}` and Run will wait for signal from that channel.
|
||
|
// Returns error only if worker fails to start.
|
||
|
func (aw *AggregatedWorker) Run(interruptCh <-chan interface{}) error {
|
||
|
if err := aw.Start(); err != nil {
|
||
|
return err
|
||
|
}
|
||
|
select {
|
||
|
case s := <-interruptCh:
|
||
|
aw.logger.Info("Worker has been stopped.", "Signal", s)
|
||
|
aw.Stop()
|
||
|
case <-aw.stopC:
|
||
|
}
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
// Stop the worker.
|
||
|
func (aw *AggregatedWorker) Stop() {
|
||
|
close(aw.stopC)
|
||
|
|
||
|
if !util.IsInterfaceNil(aw.workflowWorker) {
|
||
|
aw.workflowWorker.Stop()
|
||
|
}
|
||
|
if !util.IsInterfaceNil(aw.activityWorker) {
|
||
|
aw.activityWorker.Stop()
|
||
|
}
|
||
|
if !util.IsInterfaceNil(aw.sessionWorker) {
|
||
|
aw.sessionWorker.Stop()
|
||
|
}
|
||
|
|
||
|
aw.logger.Info("Stopped Worker")
|
||
|
}
|
||
|
|
||
|
// WorkflowReplayer is used to replay workflow code from an event history
|
||
|
type WorkflowReplayer struct {
|
||
|
registry *registry
|
||
|
dataConverter converter.DataConverter
|
||
|
}
|
||
|
|
||
|
// WorkflowReplayerOptions are options for creating a workflow replayer.
|
||
|
type WorkflowReplayerOptions struct {
|
||
|
// Optional custom data converter to provide for replay. If not set, the
|
||
|
// default converter is used.
|
||
|
DataConverter converter.DataConverter
|
||
|
}
|
||
|
|
||
|
// NewWorkflowReplayer creates an instance of the WorkflowReplayer.
|
||
|
func NewWorkflowReplayer(options WorkflowReplayerOptions) (*WorkflowReplayer, error) {
|
||
|
return &WorkflowReplayer{
|
||
|
registry: newRegistry(),
|
||
|
dataConverter: options.DataConverter,
|
||
|
}, nil
|
||
|
}
|
||
|
|
||
|
// RegisterWorkflow registers workflow function to replay
|
||
|
func (aw *WorkflowReplayer) RegisterWorkflow(w interface{}) {
|
||
|
aw.registry.RegisterWorkflow(w)
|
||
|
}
|
||
|
|
||
|
// RegisterWorkflowWithOptions registers workflow function with custom workflow name to replay
|
||
|
func (aw *WorkflowReplayer) RegisterWorkflowWithOptions(w interface{}, options RegisterWorkflowOptions) {
|
||
|
aw.registry.RegisterWorkflowWithOptions(w, options)
|
||
|
}
|
||
|
|
||
|
// ReplayWorkflowHistory executes a single workflow task for the given history.
|
||
|
// Use for testing the backwards compatibility of code changes and troubleshooting workflows in a debugger.
|
||
|
// The logger is an optional parameter. Defaults to the noop logger.
|
||
|
func (aw *WorkflowReplayer) ReplayWorkflowHistory(logger log.Logger, history *historypb.History) error {
|
||
|
if logger == nil {
|
||
|
logger = ilog.NewDefaultLogger()
|
||
|
}
|
||
|
|
||
|
controller := gomock.NewController(ilog.NewTestReporter(logger))
|
||
|
service := workflowservicemock.NewMockWorkflowServiceClient(controller)
|
||
|
|
||
|
return aw.replayWorkflowHistory(logger, service, ReplayNamespace, history)
|
||
|
}
|
||
|
|
||
|
// ReplayWorkflowHistoryFromJSONFile executes a single workflow task for the given json history file.
|
||
|
// Use for testing the backwards compatibility of code changes and troubleshooting workflows in a debugger.
|
||
|
// The logger is an optional parameter. Defaults to the noop logger.
|
||
|
func (aw *WorkflowReplayer) ReplayWorkflowHistoryFromJSONFile(logger log.Logger, jsonfileName string) error {
|
||
|
return aw.ReplayPartialWorkflowHistoryFromJSONFile(logger, jsonfileName, 0)
|
||
|
}
|
||
|
|
||
|
// ReplayPartialWorkflowHistoryFromJSONFile executes a single workflow task for the given json history file upto provided
|
||
|
// lastEventID(inclusive).
|
||
|
// Use for testing the backwards compatibility of code changes and troubleshooting workflows in a debugger.
|
||
|
// The logger is an optional parameter. Defaults to the noop logger.
|
||
|
func (aw *WorkflowReplayer) ReplayPartialWorkflowHistoryFromJSONFile(logger log.Logger, jsonfileName string, lastEventID int64) error {
|
||
|
history, err := extractHistoryFromFile(jsonfileName, lastEventID)
|
||
|
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
if logger == nil {
|
||
|
logger = ilog.NewDefaultLogger()
|
||
|
}
|
||
|
|
||
|
controller := gomock.NewController(ilog.NewTestReporter(logger))
|
||
|
service := workflowservicemock.NewMockWorkflowServiceClient(controller)
|
||
|
|
||
|
return aw.replayWorkflowHistory(logger, service, ReplayNamespace, history)
|
||
|
}
|
||
|
|
||
|
// ReplayWorkflowExecution replays workflow execution loading it from Temporal service.
|
||
|
func (aw *WorkflowReplayer) ReplayWorkflowExecution(ctx context.Context, service workflowservice.WorkflowServiceClient, logger log.Logger, namespace string, execution WorkflowExecution) error {
|
||
|
if logger == nil {
|
||
|
logger = ilog.NewDefaultLogger()
|
||
|
}
|
||
|
|
||
|
sharedExecution := &commonpb.WorkflowExecution{
|
||
|
RunId: execution.RunID,
|
||
|
WorkflowId: execution.ID,
|
||
|
}
|
||
|
request := &workflowservice.GetWorkflowExecutionHistoryRequest{
|
||
|
Namespace: namespace,
|
||
|
Execution: sharedExecution,
|
||
|
}
|
||
|
hResponse, err := service.GetWorkflowExecutionHistory(ctx, request)
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
if hResponse.RawHistory != nil {
|
||
|
history, err := serializer.DeserializeBlobDataToHistoryEvents(hResponse.RawHistory, enumspb.HISTORY_EVENT_FILTER_TYPE_ALL_EVENT)
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
hResponse.History = history
|
||
|
}
|
||
|
|
||
|
return aw.replayWorkflowHistory(logger, service, namespace, hResponse.History)
|
||
|
}
|
||
|
|
||
|
func (aw *WorkflowReplayer) replayWorkflowHistory(logger log.Logger, service workflowservice.WorkflowServiceClient, namespace string, history *historypb.History) error {
|
||
|
taskQueue := "ReplayTaskQueue"
|
||
|
events := history.Events
|
||
|
if events == nil {
|
||
|
return errors.New("empty events")
|
||
|
}
|
||
|
if len(events) < 3 {
|
||
|
return errors.New("at least 3 events expected in the history")
|
||
|
}
|
||
|
first := events[0]
|
||
|
if first.GetEventType() != enumspb.EVENT_TYPE_WORKFLOW_EXECUTION_STARTED {
|
||
|
return errors.New("first event is not WorkflowExecutionStarted")
|
||
|
}
|
||
|
last := events[len(events)-1]
|
||
|
|
||
|
attr := first.GetWorkflowExecutionStartedEventAttributes()
|
||
|
if attr == nil {
|
||
|
return errors.New("corrupted WorkflowExecutionStarted")
|
||
|
}
|
||
|
workflowType := attr.WorkflowType
|
||
|
execution := &commonpb.WorkflowExecution{
|
||
|
RunId: uuid.NewRandom().String(),
|
||
|
WorkflowId: "ReplayId",
|
||
|
}
|
||
|
if first.GetWorkflowExecutionStartedEventAttributes().GetOriginalExecutionRunId() != "" {
|
||
|
execution.RunId = first.GetWorkflowExecutionStartedEventAttributes().GetOriginalExecutionRunId()
|
||
|
}
|
||
|
|
||
|
task := &workflowservice.PollWorkflowTaskQueueResponse{
|
||
|
Attempt: 1,
|
||
|
TaskToken: []byte("ReplayTaskToken"),
|
||
|
WorkflowType: workflowType,
|
||
|
WorkflowExecution: execution,
|
||
|
History: history,
|
||
|
PreviousStartedEventId: math.MaxInt64,
|
||
|
}
|
||
|
|
||
|
iterator := &historyIteratorImpl{
|
||
|
nextPageToken: task.NextPageToken,
|
||
|
execution: task.WorkflowExecution,
|
||
|
namespace: ReplayNamespace,
|
||
|
service: service,
|
||
|
maxEventID: task.GetStartedEventId(),
|
||
|
taskQueue: taskQueue,
|
||
|
}
|
||
|
cache := NewWorkerCache()
|
||
|
params := workerExecutionParameters{
|
||
|
Namespace: namespace,
|
||
|
TaskQueue: taskQueue,
|
||
|
Identity: "replayID",
|
||
|
Logger: logger,
|
||
|
cache: cache,
|
||
|
DataConverter: aw.dataConverter,
|
||
|
}
|
||
|
taskHandler := newWorkflowTaskHandler(params, nil, aw.registry)
|
||
|
resp, err := taskHandler.ProcessWorkflowTask(&workflowTask{task: task, historyIterator: iterator}, nil)
|
||
|
if err != nil {
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
if failedReq, ok := resp.(*workflowservice.RespondWorkflowTaskFailedRequest); ok {
|
||
|
return fmt.Errorf("replay workflow failed with failure: %v", failedReq.GetFailure())
|
||
|
}
|
||
|
|
||
|
if last.GetEventType() != enumspb.EVENT_TYPE_WORKFLOW_EXECUTION_COMPLETED && last.GetEventType() != enumspb.EVENT_TYPE_WORKFLOW_EXECUTION_CONTINUED_AS_NEW {
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
if resp != nil {
|
||
|
completeReq, ok := resp.(*workflowservice.RespondWorkflowTaskCompletedRequest)
|
||
|
if ok {
|
||
|
for _, d := range completeReq.Commands {
|
||
|
if d.GetCommandType() == enumspb.COMMAND_TYPE_CONTINUE_AS_NEW_WORKFLOW_EXECUTION {
|
||
|
if last.GetEventType() == enumspb.EVENT_TYPE_WORKFLOW_EXECUTION_CONTINUED_AS_NEW {
|
||
|
inputA := d.GetContinueAsNewWorkflowExecutionCommandAttributes().GetInput()
|
||
|
inputB := last.GetWorkflowExecutionContinuedAsNewEventAttributes().GetInput()
|
||
|
if proto.Equal(inputA, inputB) {
|
||
|
return nil
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
if d.GetCommandType() == enumspb.COMMAND_TYPE_COMPLETE_WORKFLOW_EXECUTION {
|
||
|
if last.GetEventType() == enumspb.EVENT_TYPE_WORKFLOW_EXECUTION_COMPLETED {
|
||
|
resultA := last.GetWorkflowExecutionCompletedEventAttributes().GetResult()
|
||
|
resultB := d.GetCompleteWorkflowExecutionCommandAttributes().GetResult()
|
||
|
if proto.Equal(resultA, resultB) {
|
||
|
return nil
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
return fmt.Errorf("replay workflow doesn't return the same result as the last event, resp: %v, last: %v", resp, last)
|
||
|
}
|
||
|
|
||
|
func extractHistoryFromFile(jsonfileName string, lastEventID int64) (*historypb.History, error) {
|
||
|
reader, err := os.Open(jsonfileName)
|
||
|
if err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
|
||
|
var deserializedHistory historypb.History
|
||
|
err = jsonpb.Unmarshal(reader, &deserializedHistory)
|
||
|
|
||
|
if err != nil {
|
||
|
return nil, err
|
||
|
}
|
||
|
|
||
|
if lastEventID <= 0 {
|
||
|
return &deserializedHistory, nil
|
||
|
}
|
||
|
|
||
|
// Caller is potentially asking for subset of history instead of all history events
|
||
|
var events []*historypb.HistoryEvent
|
||
|
for _, event := range deserializedHistory.Events {
|
||
|
events = append(events, event)
|
||
|
if event.GetEventId() == lastEventID {
|
||
|
// Copy history up to last event (inclusive)
|
||
|
break
|
||
|
}
|
||
|
}
|
||
|
history := &historypb.History{Events: events}
|
||
|
|
||
|
return history, nil
|
||
|
}
|
||
|
|
||
|
// NewAggregatedWorker returns an instance to manage both activity and workflow workers
|
||
|
func NewAggregatedWorker(client *WorkflowClient, taskQueue string, options WorkerOptions) *AggregatedWorker {
|
||
|
setClientDefaults(client)
|
||
|
setWorkerOptionsDefaults(&options)
|
||
|
ctx := options.BackgroundActivityContext
|
||
|
if ctx == nil {
|
||
|
ctx = context.Background()
|
||
|
}
|
||
|
backgroundActivityContext, backgroundActivityContextCancel := context.WithCancel(ctx)
|
||
|
|
||
|
cache := NewWorkerCache()
|
||
|
workerParams := workerExecutionParameters{
|
||
|
Namespace: client.namespace,
|
||
|
TaskQueue: taskQueue,
|
||
|
ConcurrentActivityExecutionSize: options.MaxConcurrentActivityExecutionSize,
|
||
|
WorkerActivitiesPerSecond: options.WorkerActivitiesPerSecond,
|
||
|
MaxConcurrentActivityTaskQueuePollers: options.MaxConcurrentActivityTaskPollers,
|
||
|
ConcurrentLocalActivityExecutionSize: options.MaxConcurrentLocalActivityExecutionSize,
|
||
|
WorkerLocalActivitiesPerSecond: options.WorkerLocalActivitiesPerSecond,
|
||
|
ConcurrentWorkflowTaskExecutionSize: options.MaxConcurrentWorkflowTaskExecutionSize,
|
||
|
MaxConcurrentWorkflowTaskQueuePollers: options.MaxConcurrentWorkflowTaskPollers,
|
||
|
Identity: client.identity,
|
||
|
MetricsHandler: client.metricsHandler,
|
||
|
Logger: client.logger,
|
||
|
EnableLoggingInReplay: options.EnableLoggingInReplay,
|
||
|
UserContext: backgroundActivityContext,
|
||
|
UserContextCancel: backgroundActivityContextCancel,
|
||
|
StickyScheduleToStartTimeout: options.StickyScheduleToStartTimeout,
|
||
|
TaskQueueActivitiesPerSecond: options.TaskQueueActivitiesPerSecond,
|
||
|
WorkflowPanicPolicy: options.WorkflowPanicPolicy,
|
||
|
DataConverter: client.dataConverter,
|
||
|
WorkerStopTimeout: options.WorkerStopTimeout,
|
||
|
ContextPropagators: client.contextPropagators,
|
||
|
DeadlockDetectionTimeout: options.DeadlockDetectionTimeout,
|
||
|
DefaultHeartbeatThrottleInterval: options.DefaultHeartbeatThrottleInterval,
|
||
|
MaxHeartbeatThrottleInterval: options.MaxHeartbeatThrottleInterval,
|
||
|
cache: cache,
|
||
|
}
|
||
|
|
||
|
if options.Identity != "" {
|
||
|
workerParams.Identity = options.Identity
|
||
|
}
|
||
|
|
||
|
ensureRequiredParams(&workerParams)
|
||
|
workerParams.Logger = log.With(workerParams.Logger,
|
||
|
tagNamespace, client.namespace,
|
||
|
tagTaskQueue, taskQueue,
|
||
|
tagWorkerID, workerParams.Identity,
|
||
|
)
|
||
|
|
||
|
processTestTags(&options, &workerParams)
|
||
|
|
||
|
// worker specific registry
|
||
|
registry := newRegistry()
|
||
|
// Build set of interceptors using the applicable client ones first (being
|
||
|
// careful not to append to the existing slice)
|
||
|
registry.interceptors = make([]WorkerInterceptor, 0, len(client.workerInterceptors)+len(options.Interceptors))
|
||
|
registry.interceptors = append(append(registry.interceptors, client.workerInterceptors...), options.Interceptors...)
|
||
|
|
||
|
// workflow factory.
|
||
|
var workflowWorker *workflowWorker
|
||
|
if !options.DisableWorkflowWorker {
|
||
|
testTags := getTestTags(options.BackgroundActivityContext)
|
||
|
if len(testTags) > 0 {
|
||
|
workflowWorker = newWorkflowWorkerWithPressurePoints(client.workflowService, workerParams, testTags, registry)
|
||
|
} else {
|
||
|
workflowWorker = newWorkflowWorker(client.workflowService, workerParams, nil, registry)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// activity types.
|
||
|
var activityWorker *activityWorker
|
||
|
if !options.LocalActivityWorkerOnly {
|
||
|
activityWorker = newActivityWorker(client.workflowService, workerParams, nil, registry, nil)
|
||
|
}
|
||
|
|
||
|
var sessionWorker *sessionWorker
|
||
|
if options.EnableSessionWorker && !options.LocalActivityWorkerOnly {
|
||
|
sessionWorker = newSessionWorker(client.workflowService, workerParams, nil, registry, options.MaxConcurrentSessionExecutionSize)
|
||
|
registry.RegisterActivityWithOptions(sessionCreationActivity, RegisterActivityOptions{
|
||
|
Name: sessionCreationActivityName,
|
||
|
})
|
||
|
registry.RegisterActivityWithOptions(sessionCompletionActivity, RegisterActivityOptions{
|
||
|
Name: sessionCompletionActivityName,
|
||
|
})
|
||
|
}
|
||
|
|
||
|
return &AggregatedWorker{
|
||
|
workflowWorker: workflowWorker,
|
||
|
activityWorker: activityWorker,
|
||
|
sessionWorker: sessionWorker,
|
||
|
logger: workerParams.Logger,
|
||
|
registry: registry,
|
||
|
stopC: make(chan struct{}),
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func processTestTags(wOptions *WorkerOptions, ep *workerExecutionParameters) {
|
||
|
testTags := getTestTags(wOptions.BackgroundActivityContext)
|
||
|
if testTags != nil {
|
||
|
if paramsOverride, ok := testTags[workerOptionsConfig]; ok {
|
||
|
for key, val := range paramsOverride {
|
||
|
switch key {
|
||
|
case workerOptionsConfigConcurrentPollRoutineSize:
|
||
|
if size, err := strconv.Atoi(val); err == nil {
|
||
|
ep.MaxConcurrentActivityTaskQueuePollers = size
|
||
|
ep.MaxConcurrentWorkflowTaskQueuePollers = size
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
func isWorkflowContext(inType reflect.Type) bool {
|
||
|
// NOTE: We don't expect any one to derive from workflow context.
|
||
|
return inType == reflect.TypeOf((*Context)(nil)).Elem()
|
||
|
}
|
||
|
|
||
|
func isValidResultType(inType reflect.Type) bool {
|
||
|
// https://golang.org/pkg/reflect/#Kind
|
||
|
switch inType.Kind() {
|
||
|
case reflect.Func, reflect.Chan, reflect.UnsafePointer:
|
||
|
return false
|
||
|
}
|
||
|
|
||
|
return true
|
||
|
}
|
||
|
|
||
|
func isError(inType reflect.Type) bool {
|
||
|
errorElem := reflect.TypeOf((*error)(nil)).Elem()
|
||
|
return inType != nil && inType.Implements(errorElem)
|
||
|
}
|
||
|
|
||
|
func getFunctionName(i interface{}) (name string, isMethod bool) {
|
||
|
if fullName, ok := i.(string); ok {
|
||
|
return fullName, false
|
||
|
}
|
||
|
fullName := runtime.FuncForPC(reflect.ValueOf(i).Pointer()).Name()
|
||
|
// Full function name that has a struct pointer receiver has the following format
|
||
|
// <prefix>.(*<type>).<function>
|
||
|
isMethod = strings.ContainsAny(fullName, "*")
|
||
|
elements := strings.Split(fullName, ".")
|
||
|
shortName := elements[len(elements)-1]
|
||
|
// This allows to call activities by method pointer
|
||
|
// Compiler adds -fm suffix to a function name which has a receiver
|
||
|
// Note that this works even if struct pointer used to get the function is nil
|
||
|
// It is possible because nil receivers are allowed.
|
||
|
// For example:
|
||
|
// var a *Activities
|
||
|
// ExecuteActivity(ctx, a.Foo)
|
||
|
// will call this function which is going to return "Foo"
|
||
|
return strings.TrimSuffix(shortName, "-fm"), isMethod
|
||
|
}
|
||
|
|
||
|
func getActivityFunctionName(r *registry, i interface{}) string {
|
||
|
result, _ := getFunctionName(i)
|
||
|
if alias, ok := r.getActivityAlias(result); ok {
|
||
|
result = alias
|
||
|
}
|
||
|
return result
|
||
|
}
|
||
|
|
||
|
func getWorkflowFunctionName(r *registry, workflowFunc interface{}) (string, error) {
|
||
|
fnName := ""
|
||
|
fType := reflect.TypeOf(workflowFunc)
|
||
|
switch getKind(fType) {
|
||
|
case reflect.String:
|
||
|
fnName = reflect.ValueOf(workflowFunc).String()
|
||
|
case reflect.Func:
|
||
|
fnName, _ = getFunctionName(workflowFunc)
|
||
|
if alias, ok := r.getWorkflowAlias(fnName); ok {
|
||
|
fnName = alias
|
||
|
}
|
||
|
default:
|
||
|
return "", fmt.Errorf("invalid type 'workflowFunc' parameter provided, it can be either worker function or function name: %v", workflowFunc)
|
||
|
}
|
||
|
|
||
|
return fnName, nil
|
||
|
}
|
||
|
|
||
|
func getReadOnlyChannel(c chan struct{}) <-chan struct{} {
|
||
|
return c
|
||
|
}
|
||
|
|
||
|
func setWorkerOptionsDefaults(options *WorkerOptions) {
|
||
|
if options.MaxConcurrentActivityExecutionSize == 0 {
|
||
|
options.MaxConcurrentActivityExecutionSize = defaultMaxConcurrentActivityExecutionSize
|
||
|
}
|
||
|
if options.WorkerActivitiesPerSecond == 0 {
|
||
|
options.WorkerActivitiesPerSecond = defaultWorkerActivitiesPerSecond
|
||
|
}
|
||
|
if options.MaxConcurrentActivityTaskPollers <= 0 {
|
||
|
options.MaxConcurrentActivityTaskPollers = defaultConcurrentPollRoutineSize
|
||
|
}
|
||
|
if options.MaxConcurrentWorkflowTaskExecutionSize == 0 {
|
||
|
options.MaxConcurrentWorkflowTaskExecutionSize = defaultMaxConcurrentTaskExecutionSize
|
||
|
}
|
||
|
if options.MaxConcurrentWorkflowTaskPollers <= 0 {
|
||
|
options.MaxConcurrentWorkflowTaskPollers = defaultConcurrentPollRoutineSize
|
||
|
}
|
||
|
if options.MaxConcurrentLocalActivityExecutionSize == 0 {
|
||
|
options.MaxConcurrentLocalActivityExecutionSize = defaultMaxConcurrentLocalActivityExecutionSize
|
||
|
}
|
||
|
if options.WorkerLocalActivitiesPerSecond == 0 {
|
||
|
options.WorkerLocalActivitiesPerSecond = defaultWorkerLocalActivitiesPerSecond
|
||
|
}
|
||
|
if options.TaskQueueActivitiesPerSecond == 0 {
|
||
|
options.TaskQueueActivitiesPerSecond = defaultTaskQueueActivitiesPerSecond
|
||
|
}
|
||
|
if options.StickyScheduleToStartTimeout.Seconds() == 0 {
|
||
|
options.StickyScheduleToStartTimeout = stickyWorkflowTaskScheduleToStartTimeoutSeconds * time.Second
|
||
|
}
|
||
|
if options.MaxConcurrentSessionExecutionSize == 0 {
|
||
|
options.MaxConcurrentSessionExecutionSize = defaultMaxConcurrentSessionExecutionSize
|
||
|
}
|
||
|
if options.DeadlockDetectionTimeout == 0 {
|
||
|
if debugMode {
|
||
|
options.DeadlockDetectionTimeout = unlimitedDeadlockDetectionTimeout
|
||
|
} else {
|
||
|
options.DeadlockDetectionTimeout = defaultDeadlockDetectionTimeout
|
||
|
}
|
||
|
}
|
||
|
if options.DefaultHeartbeatThrottleInterval == 0 {
|
||
|
options.DefaultHeartbeatThrottleInterval = defaultDefaultHeartbeatThrottleInterval
|
||
|
}
|
||
|
if options.MaxHeartbeatThrottleInterval == 0 {
|
||
|
options.MaxHeartbeatThrottleInterval = defaultMaxHeartbeatThrottleInterval
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// setClientDefaults should be needed only in unit tests.
|
||
|
func setClientDefaults(client *WorkflowClient) {
|
||
|
if client.dataConverter == nil {
|
||
|
client.dataConverter = converter.GetDefaultDataConverter()
|
||
|
}
|
||
|
if client.namespace == "" {
|
||
|
client.namespace = DefaultNamespace
|
||
|
}
|
||
|
if client.metricsHandler == nil {
|
||
|
client.metricsHandler = metrics.NopHandler
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// getTestTags returns the test tags in the context.
|
||
|
func getTestTags(ctx context.Context) map[string]map[string]string {
|
||
|
if ctx != nil {
|
||
|
env := ctx.Value(testTagsContextKey)
|
||
|
if env != nil {
|
||
|
return env.(map[string]map[string]string)
|
||
|
}
|
||
|
}
|
||
|
return nil
|
||
|
}
|
||
|
|
||
|
// Same as executeFunction but injects the context as the first parameter if the
|
||
|
// function takes it (regardless of existing parameters).
|
||
|
func executeFunctionWithContext(ctx context.Context, fn interface{}, args []interface{}) (interface{}, error) {
|
||
|
if fnType := reflect.TypeOf(fn); fnType.NumIn() > 0 && isActivityContext(fnType.In(0)) {
|
||
|
args = append([]interface{}{ctx}, args...)
|
||
|
}
|
||
|
return executeFunction(fn, args)
|
||
|
}
|
||
|
|
||
|
// Executes function and ensures that there is always 1 or 2 results and second
|
||
|
// result is error.
|
||
|
func executeFunction(fn interface{}, args []interface{}) (interface{}, error) {
|
||
|
fnValue := reflect.ValueOf(fn)
|
||
|
reflectArgs := make([]reflect.Value, len(args))
|
||
|
for i, arg := range args {
|
||
|
// If the argument is nil, use zero value
|
||
|
if arg == nil {
|
||
|
reflectArgs[i] = reflect.New(fnValue.Type().In(i)).Elem()
|
||
|
} else {
|
||
|
reflectArgs[i] = reflect.ValueOf(arg)
|
||
|
}
|
||
|
}
|
||
|
retValues := fnValue.Call(reflectArgs)
|
||
|
|
||
|
// Expect either error or (result, error)
|
||
|
if len(retValues) == 0 || len(retValues) > 2 {
|
||
|
fnName, _ := getFunctionName(fn)
|
||
|
return nil, fmt.Errorf(
|
||
|
"the function: %v signature returns %d results, it is expecting to return either error or (result, error)",
|
||
|
fnName, len(retValues))
|
||
|
}
|
||
|
// Convert error
|
||
|
var err error
|
||
|
if errResult := retValues[len(retValues)-1].Interface(); errResult != nil {
|
||
|
var ok bool
|
||
|
if err, ok = errResult.(error); !ok {
|
||
|
return nil, fmt.Errorf(
|
||
|
"failed to serialize error result as it is not of error interface: %v",
|
||
|
errResult)
|
||
|
}
|
||
|
}
|
||
|
// If there are two results, convert the first only if it's not a nil pointer
|
||
|
var res interface{}
|
||
|
if len(retValues) > 1 && (retValues[0].Kind() != reflect.Ptr || !retValues[0].IsNil()) {
|
||
|
res = retValues[0].Interface()
|
||
|
}
|
||
|
return res, err
|
||
|
}
|