peridot/infrastructure/dev-helm/temporal/values.yaml
2022-07-07 22:13:21 +02:00

408 lines
9.8 KiB
YAML

nameOverride: ""
fullnameOverride: ""
# Chart debug mode
# (eg. disable helm hook delete policy)
debug: false
# Custom Service account management
serviceAccount:
# Whether to create service account or not
create: false
# Name of the service account, default: temporal.fullname
name:
# extraAnnotations would let users add additional annotations
extraAnnotations:
server:
enabled: true
sidecarContainers:
image:
repository: temporalio/server
tag: 1.15.2
pullPolicy: IfNotPresent
# Global default settings (can be overridden per service)
replicaCount: 1
metrics:
# Annotate pods directly with Prometheus annotations.
# Use this if you installed Prometheus from a Helm chart.
annotations:
enabled: true
# Enable Prometheus ServiceMonitor
# Use this if you installed the Prometheus Operator (https://github.com/coreos/prometheus-operator).
serviceMonitor:
enabled: false
interval: 30s
# Set additional lables to all the ServiceMonitor resources
additionalLabels: {}
# label1: value1
# label2: value2
# Set Prometheus metric_relabel_configs via ServiceMonitor
# Use metricRelabelings to adjust metric and label names as needed
metricRelabelings: []
# - action: replace
# sourceLabels:
# - exported_namespace
# targetLabel: temporal_namespace
# - action: replace
# regex: service_errors_(.+)
# replacement: ${1}
# sourceLabels:
# - __name__
# targetLabel: temporal_error_kind
# - action: replace
# regex: service_errors_.+
# replacement: temporal_service_errors
# sourceLabels:
# - __name__
# targetLabel: __name__
prometheus:
timerType: histogram
podAnnotations: {}
resources:
{}
# We usually recommend not to specify default resources and to leave this as a conscious
# choice for the user. This also increases chances charts run on environments with little
# resources, such as Minikube. If you do want to specify resources, uncomment the following
# lines, adjust them as necessary, and remove the curly braces after 'resources:'.
# limits:
# cpu: 100m
# memory: 128Mi
# requests:
# cpu: 100m
# memory: 128Mi
nodeSelector: {}
tolerations: []
affinity: {}
additionalVolumes: []
additionalVolumeMounts: []
config:
logLevel: "debug,info"
# IMPORTANT: This value cannot be changed, once it's set.
numHistoryShards: 512
persistence:
defaultStore: default
additionalStores: {}
default:
driver: "cassandra"
cassandra:
hosts: []
# port: 9042
keyspace: "temporal"
user: "user"
password: "password"
existingSecret: ""
replicationFactor: 1
consistency:
default:
consistency: "local_quorum"
serialConsistency: "local_serial"
# datacenter: "us-east-1a"
# maxQPS: 1000
# maxConns: 2
sql:
driver: "mysql"
host: "mysql"
port: 3306
database: "temporal"
user: "temporal"
password: "temporal"
existingSecret: ""
secretName: ""
maxConns: 20
maxConnLifetime: "1h"
# connectAttributes:
# tx_isolation: 'READ-COMMITTED'
visibility:
driver: "cassandra"
cassandra:
hosts: []
# port: 9042
keyspace: "temporal_visibility"
user: "user"
password: "password"
existingSecret: ""
# datacenter: "us-east-1a"
# maxQPS: 1000
# maxConns: 2
replicationFactor: 1
consistency:
default:
consistency: "local_quorum"
serialConsistency: "local_serial"
sql:
driver: "mysql"
host: "mysql"
port: 3306
database: "temporal_visibility"
user: "temporal"
password: "temporal"
existingSecret: ""
secretName: ""
maxConns: 20
maxConnLifetime: "1h"
# connectAttributes:
# tx_isolation: 'READ-COMMITTED'
frontend:
# replicaCount: 1
service:
annotations: {} # Evaluated as template
type: ClusterIP
port: 7233
metrics:
annotations:
enabled: true
serviceMonitor: {}
# enabled: false
prometheus: {}
# timerType: histogram
podAnnotations: {}
resources: {}
nodeSelector: {}
tolerations: []
affinity: {}
history:
# replicaCount: 1
service:
# type: ClusterIP
port: 7234
metrics:
annotations:
enabled: true
serviceMonitor: {}
# enabled: false
prometheus: {}
# timerType: histogram
podAnnotations: {}
resources: {}
nodeSelector: {}
tolerations: []
affinity: {}
matching:
# replicaCount: 1
service:
# type: ClusterIP
port: 7235
metrics:
annotations:
enabled: false
serviceMonitor: {}
# enabled: false
prometheus: {}
# timerType: histogram
podAnnotations: {}
resources: {}
nodeSelector: {}
tolerations: []
affinity: {}
worker:
# replicaCount: 1
service:
# type: ClusterIP
port: 7239
metrics:
annotations:
enabled: true
serviceMonitor: {}
# enabled: false
prometheus: {}
# timerType: histogram
podAnnotations: {}
resources: {}
nodeSelector: {}
tolerations: []
affinity: {}
admintools:
enabled: true
image:
repository: temporalio/admin-tools
tag: 1.15.2
pullPolicy: IfNotPresent
service:
type: ClusterIP
port: 22
annotations: {}
nodeSelector: {}
tolerations: []
affinity: {}
web:
enabled: true
config:
# server/config.yml file content
auth:
enabled: false
routing:
default_to_namespace: # internal use only
issue_report_link: https://github.com/temporalio/web/issues/new/choose # set this field if you need to direct people to internal support forums
replicaCount: 1
image:
repository: temporalio/web
tag: 1.14.0
pullPolicy: IfNotPresent
service:
# set type to NodePort if access to web needs access from outside the cluster
# for more info see https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types
type: ClusterIP
port: 8088
annotations: {}
# loadBalancerIP:
ingress:
enabled: false
annotations: {}
# kubernetes.io/ingress.class: traefik
# ingress.kubernetes.io/ssl-redirect: "false"
# traefik.frontend.rule.type: PathPrefix
hosts:
- "/"
# - "domain.com/xyz"
# - "domain.com"
tls: []
# - secretName: chart-example-tls
# hosts:
# - chart-example.local
podAnnotations: {}
resources: {}
# We usually recommend not to specify default resources and to leave this as a conscious
# choice for the user. This also increases chances charts run on environments with little
# resources, such as Minikube. If you do want to specify resources, uncomment the following
# lines, adjust them as necessary, and remove the curly braces after 'resources:'.
# limits:
# cpu: 100m
# memory: 128Mi
# requests:
# cpu: 100m
# memory: 128Mi
nodeSelector: {}
tolerations: []
affinity: {}
schema:
setup:
enabled: true
backoffLimit: 100
update:
enabled: true
backoffLimit: 100
elasticsearch:
enabled: true
replicas: 3
persistence:
enabled: false
imageTag: 7.16.2
host: elasticsearch-master-headless
scheme: http
port: 9200
version: "v7"
logLevel: "error"
username: ""
password: ""
visibilityIndex: "temporal_visibility_v1_dev"
prometheus:
enabled: true
nodeExporter:
enabled: false
grafana:
enabled: true
replicas: 1
testFramework:
enabled: false
rbac:
create: false
pspEnabled: false
namespaced: true
dashboardProviders:
dashboardproviders.yaml:
apiVersion: 1
providers:
- name: 'default'
orgId: 1
folder: ''
type: file
disableDeletion: false
editable: true
options:
path: /var/lib/grafana/dashboards/default
datasources:
datasources.yaml:
apiVersion: 1
datasources:
- name: TemporalMetrics
type: prometheus
url: http://{{ .Release.Name }}-prometheus-server
access: proxy
isDefault: true
dashboards:
default:
frontend-github:
url: https://raw.githubusercontent.com/temporalio/temporal-dashboards/master/dashboards/frontend.json
datasource: TemporalMetrics
temporal-github:
url: https://raw.githubusercontent.com/temporalio/temporal-dashboards/master/dashboards/temporal.json
datasource: TemporalMetrics
history-github:
url: https://raw.githubusercontent.com/temporalio/temporal-dashboards/master/dashboards/history.json
datasource: TemporalMetrics
matching-github:
url: https://raw.githubusercontent.com/temporalio/temporal-dashboards/master/dashboards/matching.json
datasource: TemporalMetrics
clusteroverview-github:
url: https://raw.githubusercontent.com/temporalio/temporal-dashboards/master/dashboards/10000.json
datasource: TemporalMetrics
common-github:
url: https://raw.githubusercontent.com/temporalio/temporal-dashboards/master/dashboards/common.json
datasource: TemporalMetrics
cassandra:
enabled: true
persistence:
enabled: false
image:
repo: cassandra
tag: 3.11.3
pullPolicy: IfNotPresent
config:
cluster_size: 3
ports:
cql: 9042
num_tokens: 4
max_heap_size: 512M
heap_new_size: 128M
seed_size: 0
service:
type: ClusterIP
mysql:
enabled: false