peridot/vendor/cirello.io/dynamolock/client.go
2022-07-07 22:13:21 +02:00

1027 lines
33 KiB
Go

/*
Copyright 2015 github.com/ucirello
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
*/
package dynamolock
import (
"context"
"crypto/rand"
"errors"
"fmt"
"io/ioutil"
"log"
"math/big"
"sync"
"time"
"github.com/aws/aws-sdk-go/aws"
"github.com/aws/aws-sdk-go/service/dynamodb"
"github.com/aws/aws-sdk-go/service/dynamodb/dynamodbiface"
"github.com/aws/aws-sdk-go/service/dynamodb/expression"
)
const (
attrData = "data"
attrOwnerName = "ownerName"
attrLeaseDuration = "leaseDuration"
attrRecordVersionNumber = "recordVersionNumber"
attrIsReleased = "isReleased"
defaultBuffer = 1 * time.Second
)
var (
dataAttr = expression.Name(attrData)
ownerNameAttr = expression.Name(attrOwnerName)
leaseDurationAttr = expression.Name(attrLeaseDuration)
rvnAttr = expression.Name(attrRecordVersionNumber)
isReleasedAttr = expression.Name(attrIsReleased)
)
var isReleasedAttrVal = expression.Value("1")
// Logger defines the minimum desired logger interface for the lock client.
type Logger interface {
Println(v ...interface{})
}
// Client is a dynamoDB based distributed lock client.
type Client struct {
dynamoDB dynamodbiface.DynamoDBAPI
tableName string
partitionKeyName string
leaseDuration time.Duration
heartbeatPeriod time.Duration
ownerName string
locks sync.Map
sessionMonitorCancellations sync.Map
logger Logger
stopHeartbeat context.CancelFunc
mu sync.RWMutex
closeOnce sync.Once
closed bool
}
const (
defaultPartitionKeyName = "key"
defaultLeaseDuration = 20 * time.Second
defaultHeartbeatPeriod = 5 * time.Second
)
// New creates a new dynamoDB based distributed lock client.
func New(dynamoDB dynamodbiface.DynamoDBAPI, tableName string, opts ...ClientOption) (*Client, error) {
c := &Client{
dynamoDB: dynamoDB,
tableName: tableName,
partitionKeyName: defaultPartitionKeyName,
leaseDuration: defaultLeaseDuration,
heartbeatPeriod: defaultHeartbeatPeriod,
ownerName: randString(32),
logger: log.New(ioutil.Discard, "", 0),
stopHeartbeat: func() {},
}
for _, opt := range opts {
opt(c)
}
if c.leaseDuration < 2*c.heartbeatPeriod {
return nil, errors.New("heartbeat period must be no more than half the length of the Lease Duration, " +
"or locks might expire due to the heartbeat thread taking too long to update them (recommendation is to make it much greater, for example " +
"4+ times greater)")
}
if c.heartbeatPeriod > 0 {
ctx, cancel := context.WithCancel(context.Background())
c.stopHeartbeat = cancel
go c.heartbeat(ctx)
}
return c, nil
}
// ClientOption reconfigure the lock client creation.
type ClientOption func(*Client)
// WithPartitionKeyName defines the key name used for asserting keys uniqueness.
func WithPartitionKeyName(s string) ClientOption {
return func(c *Client) { c.partitionKeyName = s }
}
// WithOwnerName changes the owner linked to the client, and by consequence to
// locks.
func WithOwnerName(s string) ClientOption {
return func(c *Client) { c.ownerName = s }
}
// WithLeaseDuration defines how long should the lease be held.
func WithLeaseDuration(d time.Duration) ClientOption {
return func(c *Client) { c.leaseDuration = d }
}
// WithHeartbeatPeriod defines the frequency of the heartbeats. Set to zero to
// disable it. Heartbeats should have no more than half of the duration of the
// lease.
func WithHeartbeatPeriod(d time.Duration) ClientOption {
return func(c *Client) { c.heartbeatPeriod = d }
}
// DisableHeartbeat disables automatic hearbeats. Use SendHeartbeat to freshen
// up the lock.
func DisableHeartbeat() ClientOption {
return WithHeartbeatPeriod(0)
}
// WithLogger injects a logger into the client, so its internals can be
// recorded.
func WithLogger(l Logger) ClientOption {
return func(c *Client) { c.logger = l }
}
// AcquireLockOption allows to change how the lock is actually held by the
// client.
type AcquireLockOption func(*acquireLockOptions)
// WithData stores the content into the lock itself.
func WithData(b []byte) AcquireLockOption {
return func(opt *acquireLockOptions) {
opt.data = b
}
}
// ReplaceData will force the new content to be stored in the key.
func ReplaceData() AcquireLockOption {
return func(opt *acquireLockOptions) {
opt.replaceData = true
}
}
// FailIfLocked will not retry to acquire the lock, instead returning.
func FailIfLocked() AcquireLockOption {
return func(opt *acquireLockOptions) {
opt.failIfLocked = true
}
}
// WithDeleteLockOnRelease defines whether or not the lock should be deleted
// when Close() is called on the resulting LockItem will force the new content
// to be stored in the key.
func WithDeleteLockOnRelease() AcquireLockOption {
return func(opt *acquireLockOptions) {
opt.deleteLockOnRelease = true
}
}
// WithRefreshPeriod defines how long to wait before trying to get the lock
// again (if set to 10 seconds, for example, it would attempt to do so every 10
// seconds).
func WithRefreshPeriod(d time.Duration) AcquireLockOption {
return func(opt *acquireLockOptions) {
opt.refreshPeriod = d
}
}
// WithAdditionalTimeToWaitForLock defines how long to wait in addition to the
// lease duration (if set to 10 minutes, this will try to acquire a lock for at
// least 10 minutes before giving up and returning an error).
func WithAdditionalTimeToWaitForLock(d time.Duration) AcquireLockOption {
return func(opt *acquireLockOptions) {
opt.additionalTimeToWaitForLock = d
}
}
// WithAdditionalAttributes stores some additional attributes with each lock.
// This can be used to add any arbitrary parameters to each lock row.
func WithAdditionalAttributes(attr map[string]*dynamodb.AttributeValue) AcquireLockOption {
return func(opt *acquireLockOptions) {
opt.additionalAttributes = attr
}
}
// WithSessionMonitor registers a callback that is triggered if the lock is
// about to expire.
//
// The purpose of this construct is to provide two abilities: provide
// the ability to determine if the lock is about to expire, and run a
// user-provided callback when the lock is about to expire. The advantage
// this provides is notification that your lock is about to expire before it
// is actually expired, and in case of leader election will help in
// preventing that there are no two leaders present simultaneously.
//
// If due to any reason heartbeating is unsuccessful for a configurable
// period of time, your lock enters into a phase known as "danger zone." It
// is during this "danger zone" that the callback will be run.
//
// Bear in mind that the callback may be null. In this
// case, no callback will be run upon the lock entering the "danger zone";
// yet, one can still make use of the Lock.IsAlmostExpired() call.
// Furthermore, non-null callbacks can only ever be executed once in a
// lock's lifetime. Independent of whether or not a callback is run, the
// client will attempt to heartbeat the lock until the lock is released or
// obtained by someone else.
//
// Consider an example which uses this mechanism for leader election. One
// way to make use of this SessionMonitor is to register a callback that
// kills the instance in case the leader's lock enters the danger zone:
func WithSessionMonitor(safeTime time.Duration, callback func()) AcquireLockOption {
return func(opt *acquireLockOptions) {
opt.sessionMonitor = &sessionMonitor{
safeTime: safeTime,
callback: callback,
}
}
}
// AcquireLock holds the defined lock.
func (c *Client) AcquireLock(key string, opts ...AcquireLockOption) (*Lock, error) {
return c.AcquireLockWithContext(context.Background(), key, opts...)
}
// AcquireLockWithContext holds the defined lock. The given context is passed
// down to the underlying dynamoDB call.
func (c *Client) AcquireLockWithContext(ctx context.Context, key string, opts ...AcquireLockOption) (*Lock, error) {
if c.isClosed() {
return nil, ErrClientClosed
}
req := &acquireLockOptions{
partitionKey: key,
}
for _, opt := range opts {
opt(req)
}
return c.acquireLock(ctx, req)
}
func (c *Client) acquireLock(ctx context.Context, opt *acquireLockOptions) (*Lock, error) {
// Hold the read lock when acquiring locks. This prevents us from
// acquiring a lock while the Client is being closed as we hold the
// write lock during close.
c.mu.RLock()
defer c.mu.RUnlock()
if c.closed {
return nil, ErrClientClosed
}
attrs := opt.additionalAttributes
contains := func(ks ...string) bool {
for _, k := range ks {
if _, ok := attrs[k]; ok {
return true
}
}
return false
}
if contains(c.partitionKeyName, attrOwnerName, attrLeaseDuration,
attrRecordVersionNumber, attrData) {
return nil, fmt.Errorf("additional attribute cannot be one of the following types: %s, %s, %s, %s, %s",
c.partitionKeyName, attrOwnerName, attrLeaseDuration, attrRecordVersionNumber, attrData)
}
getLockOptions := getLockOptions{
partitionKeyName: opt.partitionKey,
deleteLockOnRelease: opt.deleteLockOnRelease,
sessionMonitor: opt.sessionMonitor,
start: time.Now(),
replaceData: opt.replaceData,
data: opt.data,
additionalAttributes: attrs,
failIfLocked: opt.failIfLocked,
}
getLockOptions.millisecondsToWait = defaultBuffer
if opt.additionalTimeToWaitForLock > 0 {
getLockOptions.millisecondsToWait = opt.additionalTimeToWaitForLock
}
getLockOptions.refreshPeriodDuration = defaultBuffer
if opt.refreshPeriod > 0 {
getLockOptions.refreshPeriodDuration = opt.refreshPeriod
}
for {
l, err := c.storeLock(ctx, &getLockOptions)
if err != nil {
return nil, err
} else if l != nil {
return l, nil
}
c.logger.Println("Sleeping for a refresh period of ", getLockOptions.refreshPeriodDuration)
select {
case <-ctx.Done():
return nil, ctx.Err()
case <-time.After(getLockOptions.refreshPeriodDuration):
}
}
}
func (c *Client) storeLock(ctx context.Context, getLockOptions *getLockOptions) (*Lock, error) {
c.logger.Println("Call GetItem to see if the lock for ",
c.partitionKeyName, " =", getLockOptions.partitionKeyName, " exists in the table")
existingLock, err := c.getLockFromDynamoDB(ctx, *getLockOptions)
if err != nil {
return nil, err
}
var newLockData []byte
if getLockOptions.replaceData {
newLockData = getLockOptions.data
} else if existingLock != nil {
newLockData = existingLock.data
}
if newLockData == nil {
// If there is no existing data, we write the input data to the lock.
newLockData = getLockOptions.data
}
mergedAdditionalAttributes := make(map[string]*dynamodb.AttributeValue)
for k, v := range existingLock.AdditionalAttributes() {
mergedAdditionalAttributes[k] = v
}
for k, v := range getLockOptions.additionalAttributes {
mergedAdditionalAttributes[k] = v
}
getLockOptions.additionalAttributes = mergedAdditionalAttributes
item := make(map[string]*dynamodb.AttributeValue)
for k, v := range getLockOptions.additionalAttributes {
item[k] = v
}
item[c.partitionKeyName] = &dynamodb.AttributeValue{S: aws.String(getLockOptions.partitionKeyName)}
item[attrOwnerName] = &dynamodb.AttributeValue{S: aws.String(c.ownerName)}
item[attrLeaseDuration] = &dynamodb.AttributeValue{S: aws.String(c.leaseDuration.String())}
recordVersionNumber := c.generateRecordVersionNumber()
item[attrRecordVersionNumber] = &dynamodb.AttributeValue{S: aws.String(recordVersionNumber)}
if newLockData != nil {
item[attrData] = &dynamodb.AttributeValue{B: newLockData}
}
//if the existing lock does not exist or exists and is released
if existingLock == nil || existingLock.isReleased {
l, err := c.upsertAndMonitorNewOrReleasedLock(
ctx,
getLockOptions.additionalAttributes,
getLockOptions.partitionKeyName,
getLockOptions.deleteLockOnRelease,
newLockData,
item,
recordVersionNumber,
getLockOptions.sessionMonitor)
if err != nil && isLockNotGrantedError(err) {
return nil, nil
}
return l, err
}
// we know that we didnt enter the if block above because it returns at the end.
// we also know that the existingLock.isPresent() is true
if getLockOptions.lockTryingToBeAcquired == nil {
//this branch of logic only happens once, in the first iteration of the while loop
//lockTryingToBeAcquired only ever gets set to non-null values after this point.
//so it is impossible to get in this
/*
* Someone else has the lock, and they have the lock for LEASE_DURATION time. At this point, we need
* to wait at least LEASE_DURATION milliseconds before we can try to acquire the lock.
*/
// If the user has set `FailIfLocked` option, exit after the first attempt to acquire the lock.
if getLockOptions.failIfLocked {
return nil, &LockNotGrantedError{msg: "Didn't acquire lock because it is locked and request is configured not to retry."}
}
getLockOptions.lockTryingToBeAcquired = existingLock
if !getLockOptions.alreadySleptOnceForOneLeasePeriod {
getLockOptions.alreadySleptOnceForOneLeasePeriod = true
getLockOptions.millisecondsToWait += existingLock.leaseDuration
}
} else if getLockOptions.lockTryingToBeAcquired.recordVersionNumber == existingLock.recordVersionNumber && getLockOptions.lockTryingToBeAcquired.isExpired() {
/* If the version numbers match, then we can acquire the lock, assuming it has already expired */
l, err := c.upsertAndMonitorExpiredLock(
ctx,
getLockOptions.additionalAttributes,
getLockOptions.partitionKeyName,
getLockOptions.deleteLockOnRelease,
existingLock, newLockData, item,
recordVersionNumber,
getLockOptions.sessionMonitor)
if err != nil && isLockNotGrantedError(err) {
return nil, nil
}
return l, err
} else if getLockOptions.lockTryingToBeAcquired.recordVersionNumber != existingLock.recordVersionNumber {
/*
* If the version number changed since we last queried the lock, then we need to update
* lockTryingToBeAcquired as the lock has been refreshed since we last checked
*/
getLockOptions.lockTryingToBeAcquired = existingLock
}
if t := time.Since(getLockOptions.start); t > getLockOptions.millisecondsToWait {
return nil, &LockNotGrantedError{
msg: "Didn't acquire lock after sleeping",
cause: &TimeoutError{Age: t},
}
}
return nil, nil
}
func (c *Client) upsertAndMonitorExpiredLock(
ctx context.Context,
additionalAttributes map[string]*dynamodb.AttributeValue,
key string,
deleteLockOnRelease bool,
existingLock *Lock,
newLockData []byte,
item map[string]*dynamodb.AttributeValue,
recordVersionNumber string,
sessionMonitor *sessionMonitor,
) (*Lock, error) {
cond := expression.And(
expression.AttributeExists(expression.Name(c.partitionKeyName)),
expression.Equal(rvnAttr, expression.Value(existingLock.recordVersionNumber)),
)
putItemExpr, _ := expression.NewBuilder().WithCondition(cond).Build()
putItemRequest := &dynamodb.PutItemInput{
Item: item,
TableName: aws.String(c.tableName),
ConditionExpression: putItemExpr.Condition(),
ExpressionAttributeNames: putItemExpr.Names(),
ExpressionAttributeValues: putItemExpr.Values(),
}
c.logger.Println("Acquiring an existing lock whose revisionVersionNumber did not change for ",
c.partitionKeyName, " partitionKeyName=", key)
return c.putLockItemAndStartSessionMonitor(
ctx, additionalAttributes, key, deleteLockOnRelease, newLockData,
recordVersionNumber, sessionMonitor, putItemRequest)
}
func (c *Client) upsertAndMonitorNewOrReleasedLock(
ctx context.Context,
additionalAttributes map[string]*dynamodb.AttributeValue,
key string,
deleteLockOnRelease bool,
newLockData []byte,
item map[string]*dynamodb.AttributeValue,
recordVersionNumber string,
sessionMonitor *sessionMonitor,
) (*Lock, error) {
cond := expression.Or(
expression.AttributeNotExists(expression.Name(c.partitionKeyName)),
expression.And(
expression.AttributeExists(expression.Name(c.partitionKeyName)),
expression.Equal(isReleasedAttr, isReleasedAttrVal),
),
)
putItemExpr, _ := expression.NewBuilder().WithCondition(cond).Build()
req := &dynamodb.PutItemInput{
Item: item,
TableName: aws.String(c.tableName),
ConditionExpression: putItemExpr.Condition(),
ExpressionAttributeNames: putItemExpr.Names(),
ExpressionAttributeValues: putItemExpr.Values(),
}
// No one has the lock, go ahead and acquire it. The person storing the
// lock into DynamoDB should err on the side of thinking the lock will
// expire sooner than it actually will, so they start counting towards
// its expiration before the Put succeeds
c.logger.Println("Acquiring a new lock or an existing yet released lock on ", c.partitionKeyName, "=", key)
return c.putLockItemAndStartSessionMonitor(ctx, additionalAttributes, key,
deleteLockOnRelease, newLockData,
recordVersionNumber, sessionMonitor, req)
}
func (c *Client) putLockItemAndStartSessionMonitor(
ctx context.Context,
additionalAttributes map[string]*dynamodb.AttributeValue,
key string,
deleteLockOnRelease bool,
newLockData []byte,
recordVersionNumber string,
sessionMonitor *sessionMonitor,
putItemRequest *dynamodb.PutItemInput) (*Lock, error) {
lastUpdatedTime := time.Now()
_, err := c.dynamoDB.PutItemWithContext(ctx, putItemRequest)
if err != nil {
return nil, parseDynamoDBError(err, "cannot store lock item: lock already acquired by other client")
}
lockItem := &Lock{
client: c,
partitionKey: key,
data: newLockData,
deleteLockOnRelease: deleteLockOnRelease,
ownerName: c.ownerName,
leaseDuration: c.leaseDuration,
lookupTime: lastUpdatedTime,
recordVersionNumber: recordVersionNumber,
additionalAttributes: additionalAttributes,
sessionMonitor: sessionMonitor,
}
c.locks.Store(lockItem.uniqueIdentifier(), lockItem)
c.tryAddSessionMonitor(lockItem.uniqueIdentifier(), lockItem)
return lockItem, nil
}
func (c *Client) getLockFromDynamoDB(ctx context.Context, opt getLockOptions) (*Lock, error) {
res, err := c.readFromDynamoDB(ctx, opt.partitionKeyName)
if err != nil {
return nil, err
}
item := res.Item
if item == nil {
return nil, nil
}
return c.createLockItem(opt, item)
}
func (c *Client) readFromDynamoDB(ctx context.Context, key string) (*dynamodb.GetItemOutput, error) {
dynamoDBKey := map[string]*dynamodb.AttributeValue{
c.partitionKeyName: {S: aws.String(key)},
}
return c.dynamoDB.GetItemWithContext(ctx, &dynamodb.GetItemInput{
ConsistentRead: aws.Bool(true),
TableName: aws.String(c.tableName),
Key: dynamoDBKey,
})
}
func (c *Client) createLockItem(opt getLockOptions, item map[string]*dynamodb.AttributeValue) (*Lock, error) {
var data []byte
if r, ok := item[attrData]; ok {
data = r.B
delete(item, attrData)
}
ownerName := item[attrOwnerName]
delete(item, attrOwnerName)
leaseDuration := item[attrLeaseDuration]
delete(item, attrLeaseDuration)
recordVersionNumber := item[attrRecordVersionNumber]
delete(item, attrRecordVersionNumber)
_, isReleased := item[attrIsReleased]
delete(item, attrIsReleased)
delete(item, c.partitionKeyName)
// The person retrieving the lock in DynamoDB should err on the side of
// not expiring the lock, so they don't start counting until after the
// call to DynamoDB succeeds
lookupTime := time.Now()
var parsedLeaseDuration time.Duration
if leaseDuration != nil {
var err error
parsedLeaseDuration, err = time.ParseDuration(aws.StringValue(leaseDuration.S))
if err != nil {
return nil, fmt.Errorf("cannot parse lease duration: %s", err)
}
}
lockItem := &Lock{
client: c,
partitionKey: opt.partitionKeyName,
data: data,
deleteLockOnRelease: opt.deleteLockOnRelease,
ownerName: aws.StringValue(ownerName.S),
leaseDuration: parsedLeaseDuration,
lookupTime: lookupTime,
recordVersionNumber: aws.StringValue(recordVersionNumber.S),
isReleased: isReleased,
additionalAttributes: item,
}
return lockItem, nil
}
func (c *Client) generateRecordVersionNumber() string {
// TODO: improve me
return randString(32)
}
var letterRunes = []rune("1234567890abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ")
func randString(n int) string {
b := make([]rune, n)
for i := range b {
// ignoring error as the only possible error is for io.ReadFull
r, _ := rand.Int(rand.Reader, big.NewInt(int64(len(letterRunes))))
b[i] = letterRunes[r.Int64()]
}
return string(b)
}
func (c *Client) heartbeat(ctx context.Context) {
c.logger.Println("starting heartbeats")
tick := time.NewTicker(c.heartbeatPeriod)
defer tick.Stop()
for range tick.C {
c.locks.Range(func(_ interface{}, value interface{}) bool {
lockItem := value.(*Lock)
if err := c.SendHeartbeat(lockItem); err != nil {
c.logger.Println("error sending heartbeat to", lockItem.partitionKey, ":", err)
}
return true
})
if ctx.Err() != nil {
c.logger.Println("client closed, stopping heartbeat")
return
}
}
}
// CreateTable prepares a DynamoDB table with the right schema for it to be used
// by this locking library. The table should be set up in advance, because it
// takes a few minutes for DynamoDB to provision a new instance. Also, if the
// table already exists, it will return an error.
func (c *Client) CreateTable(tableName string, opts ...CreateTableOption) (*dynamodb.CreateTableOutput, error) {
return c.CreateTableWithContext(context.Background(), tableName, opts...)
}
// CreateTableWithContext prepares a DynamoDB table with the right schema for it
// to be used by this locking library. The table should be set up in advance,
// because it takes a few minutes for DynamoDB to provision a new instance.
// Also, if the table already exists, it will return an error. The given context
// is passed down to the underlying dynamoDB call.
func (c *Client) CreateTableWithContext(ctx context.Context, tableName string, opts ...CreateTableOption) (*dynamodb.CreateTableOutput, error) {
if c.isClosed() {
return nil, ErrClientClosed
}
createTableOptions := &createDynamoDBTableOptions{
tableName: tableName,
billingMode: "PAY_PER_REQUEST",
partitionKeyName: defaultPartitionKeyName,
}
for _, opt := range opts {
opt(createTableOptions)
}
return c.createTable(ctx, createTableOptions)
}
// CreateTableOption is an options type for the CreateTable method in the lock
// client. This allows the user to create a DynamoDB table that is lock
// client-compatible and specify optional parameters such as the desired
// throughput and whether or not to use a sort key.
type CreateTableOption func(*createDynamoDBTableOptions)
// WithCustomPartitionKeyName changes the partition key name of the table. If
// not specified, the default "key" will be used.
func WithCustomPartitionKeyName(s string) CreateTableOption {
return func(opt *createDynamoDBTableOptions) {
opt.partitionKeyName = s
}
}
// WithTags changes the tags of the table. If not specified, the table will have empty tags.
func WithTags(tags []*dynamodb.Tag) CreateTableOption {
return func(opt *createDynamoDBTableOptions) {
opt.tags = tags
}
}
// WithProvisionedThroughput changes the billing mode of DynamoDB
// and tells DynamoDB to operate in a provisioned throughput mode instead of pay-per-request
func WithProvisionedThroughput(provisionedThroughput *dynamodb.ProvisionedThroughput) CreateTableOption {
return func(opt *createDynamoDBTableOptions) {
opt.billingMode = "PROVISIONED"
opt.provisionedThroughput = provisionedThroughput
}
}
func (c *Client) createTable(ctx context.Context, opt *createDynamoDBTableOptions) (*dynamodb.CreateTableOutput, error) {
keySchema := []*dynamodb.KeySchemaElement{
{
AttributeName: aws.String(opt.partitionKeyName),
KeyType: aws.String(dynamodb.KeyTypeHash),
},
}
attributeDefinitions := []*dynamodb.AttributeDefinition{
{
AttributeName: aws.String(opt.partitionKeyName),
AttributeType: aws.String("S"),
},
}
createTableInput := &dynamodb.CreateTableInput{
TableName: aws.String(opt.tableName),
KeySchema: keySchema,
BillingMode: aws.String(opt.billingMode),
AttributeDefinitions: attributeDefinitions,
}
if opt.provisionedThroughput != nil {
createTableInput.ProvisionedThroughput = opt.provisionedThroughput
}
if opt.tags != nil {
createTableInput.Tags = opt.tags
}
return c.dynamoDB.CreateTableWithContext(ctx, createTableInput)
}
// ReleaseLock releases the given lock if the current user still has it,
// returning true if the lock was successfully released, and false if someone
// else already stole the lock or a problem happened. Deletes the lock item if
// it is released and deleteLockItemOnClose is set.
func (c *Client) ReleaseLock(lockItem *Lock, opts ...ReleaseLockOption) (bool, error) {
return c.ReleaseLockWithContext(context.Background(), lockItem, opts...)
}
// ReleaseLockWithContext releases the given lock if the current user still has it,
// returning true if the lock was successfully released, and false if someone
// else already stole the lock or a problem happened. Deletes the lock item if
// it is released and deleteLockItemOnClose is set.
func (c *Client) ReleaseLockWithContext(ctx context.Context, lockItem *Lock, opts ...ReleaseLockOption) (bool, error) {
if c.isClosed() {
return false, ErrClientClosed
}
err := c.releaseLock(ctx, lockItem, opts...)
return err == nil, err
}
// WithDeleteLock defines whether or not to delete the lock when releasing it.
// If set to false, the lock row will continue to be in DynamoDB, but it will be
// marked as released.
func WithDeleteLock(deleteLock bool) ReleaseLockOption {
return func(opt *releaseLockOptions) {
opt.deleteLock = deleteLock
}
}
// WithDataAfterRelease is the new data to persist to the lock (only used if
// deleteLock=false.) If the data is null, then the lock client will keep the
// data as-is and not change it.
func WithDataAfterRelease(data []byte) ReleaseLockOption {
return func(opt *releaseLockOptions) {
opt.data = data
}
}
// ReleaseLockOption provides options for releasing a lock when calling the
// releaseLock() method. This class contains the options that may be configured
// during the act of releasing a lock.
type ReleaseLockOption func(*releaseLockOptions)
func ownershipLockCondition(partitionKeyName, recordVersionNumber, ownerName string) expression.ConditionBuilder {
cond := expression.And(
expression.And(
expression.AttributeExists(expression.Name(partitionKeyName)),
expression.Equal(rvnAttr, expression.Value(recordVersionNumber)),
),
expression.Equal(ownerNameAttr, expression.Value(ownerName)),
)
return cond
}
func (c *Client) releaseLock(ctx context.Context, lockItem *Lock, opts ...ReleaseLockOption) error {
options := &releaseLockOptions{
lockItem: lockItem,
}
if lockItem != nil {
options.deleteLock = lockItem.deleteLockOnRelease
}
for _, opt := range opts {
opt(options)
}
if lockItem == nil {
return ErrCannotReleaseNullLock
}
deleteLock := options.deleteLock
data := options.data
if lockItem.ownerName != c.ownerName {
return ErrOwnerMismatched
}
lockItem.semaphore.Lock()
defer lockItem.semaphore.Unlock()
lockItem.isReleased = true
c.locks.Delete(lockItem.uniqueIdentifier())
key := c.getItemKeys(lockItem)
ownershipLockCond := ownershipLockCondition(c.partitionKeyName, lockItem.recordVersionNumber, lockItem.ownerName)
if deleteLock {
err := c.deleteLock(ctx, ownershipLockCond, key)
if err != nil {
return err
}
} else {
err := c.updateLock(ctx, data, ownershipLockCond, key)
if err != nil {
return err
}
}
c.removeKillSessionMonitor(lockItem.uniqueIdentifier())
return nil
}
func (c *Client) deleteLock(ctx context.Context, ownershipLockCond expression.ConditionBuilder, key map[string]*dynamodb.AttributeValue) error {
delExpr, _ := expression.NewBuilder().WithCondition(ownershipLockCond).Build()
deleteItemRequest := &dynamodb.DeleteItemInput{
TableName: aws.String(c.tableName),
Key: key,
ConditionExpression: delExpr.Condition(),
ExpressionAttributeNames: delExpr.Names(),
ExpressionAttributeValues: delExpr.Values(),
}
_, err := c.dynamoDB.DeleteItemWithContext(ctx, deleteItemRequest)
if err != nil {
return err
}
return nil
}
func (c *Client) updateLock(ctx context.Context, data []byte, ownershipLockCond expression.ConditionBuilder, key map[string]*dynamodb.AttributeValue) error {
update := expression.Set(isReleasedAttr, isReleasedAttrVal)
if len(data) > 0 {
update = update.Set(dataAttr, expression.Value(data))
}
updateExpr, _ := expression.NewBuilder().WithUpdate(update).WithCondition(ownershipLockCond).Build()
updateItemRequest := &dynamodb.UpdateItemInput{
TableName: aws.String(c.tableName),
Key: key,
UpdateExpression: updateExpr.Update(),
ConditionExpression: updateExpr.Condition(),
ExpressionAttributeNames: updateExpr.Names(),
ExpressionAttributeValues: updateExpr.Values(),
}
_, err := c.dynamoDB.UpdateItemWithContext(ctx, updateItemRequest)
return err
}
func (c *Client) releaseAllLocks(ctx context.Context) error {
var err error
c.locks.Range(func(key interface{}, value interface{}) bool {
err = c.releaseLock(ctx, value.(*Lock))
return err == nil
})
return err
}
func (c *Client) getItemKeys(lockItem *Lock) map[string]*dynamodb.AttributeValue {
key := map[string]*dynamodb.AttributeValue{
c.partitionKeyName: {S: aws.String(lockItem.partitionKey)},
}
return key
}
// Get finds out who owns the given lock, but does not acquire the lock. It
// returns the metadata currently associated with the given lock. If the client
// currently has the lock, it will return the lock, and operations such as
// releaseLock will work. However, if the client does not have the lock, then
// operations like releaseLock will not work (after calling Get, the caller
// should check lockItem.isExpired() to figure out if it currently has the
// lock.)
func (c *Client) Get(key string) (*Lock, error) {
return c.GetWithContext(context.Background(), key)
}
// GetWithContext finds out who owns the given lock, but does not acquire the
// lock. It returns the metadata currently associated with the given lock. If
// the client currently has the lock, it will return the lock, and operations
// such as releaseLock will work. However, if the client does not have the lock,
// then operations like releaseLock will not work (after calling Get, the caller
// should check lockItem.isExpired() to figure out if it currently has the
// lock.) If the context is canceled, it is going to return the context error
// on local cache hit. The given context is passed down to the underlying
// dynamoDB call.
func (c *Client) GetWithContext(ctx context.Context, key string) (*Lock, error) {
if c.isClosed() {
return nil, ErrClientClosed
}
if err := ctx.Err(); err != nil {
return nil, err
}
getLockOption := getLockOptions{
partitionKeyName: key,
}
keyName := getLockOption.partitionKeyName
v, ok := c.locks.Load(keyName)
if ok {
return v.(*Lock), nil
}
lockItem, err := c.getLockFromDynamoDB(ctx, getLockOption)
if err != nil {
return nil, err
}
if lockItem == nil {
return &Lock{}, nil
}
lockItem.updateRVN("", time.Time{}, lockItem.leaseDuration)
return lockItem, nil
}
// ErrClientClosed reports the client cannot be used because it is already
// closed.
var ErrClientClosed = errors.New("client already closed")
func (c *Client) isClosed() bool {
c.mu.RLock()
closed := c.closed
c.mu.RUnlock()
return closed
}
// Close releases all of the locks.
func (c *Client) Close() error {
return c.CloseWithContext(context.Background())
}
// CloseWithContext releases all of the locks. The given context is passed down
// to the underlying dynamoDB calls.
func (c *Client) CloseWithContext(ctx context.Context) error {
err := ErrClientClosed
c.closeOnce.Do(func() {
// Hold the write lock for the duration of the close operation
// to prevent new locks from being acquired.
c.mu.Lock()
defer c.mu.Unlock()
err = c.releaseAllLocks(context.Background())
c.stopHeartbeat()
c.closed = true
})
return err
}
func (c *Client) tryAddSessionMonitor(lockName string, lock *Lock) {
if lock.sessionMonitor != nil && lock.sessionMonitor.callback != nil {
ctx, cancel := context.WithCancel(context.Background())
c.lockSessionMonitorChecker(ctx, lockName, lock)
c.sessionMonitorCancellations.Store(lockName, cancel)
}
}
func (c *Client) removeKillSessionMonitor(monitorName string) {
sm, ok := c.sessionMonitorCancellations.Load(monitorName)
if !ok {
return
}
if cancel, ok := sm.(func()); ok {
cancel()
} else if cancel, ok := sm.(context.CancelFunc); ok {
cancel()
}
}
func (c *Client) lockSessionMonitorChecker(ctx context.Context,
monitorName string, lock *Lock) {
go func() {
defer c.sessionMonitorCancellations.Delete(monitorName)
for {
select {
case <-ctx.Done():
return
default:
timeUntilDangerZone, err := lock.timeUntilDangerZoneEntered()
if err != nil {
c.logger.Println("cannot run session monitor because", err)
return
}
if timeUntilDangerZone <= 0 {
go lock.sessionMonitor.callback()
return
}
time.Sleep(timeUntilDangerZone)
}
}
}()
}