# Servers. conductor.jetty.server.enabled=true conductor.grpc.server.enabled=false # Database persistence model. Possible values are memory, redis, and dynomite. # If ommitted, the persistence used is memory # # memory : The data is stored in memory and lost when the server dies. Useful for testing or demo # redis : non-Dynomite based redis instance # dynomite : Dynomite cluster. Use this for HA configuration. db=dynomite # Dynomite Cluster details. # format is host:port:rack separated by semicolon workflow.dynomite.cluster.hosts={% set ns = namespace() %} {% set ns.datacenter = "us-east-" %} {% set ns.current_rack = "" %} {% set ns.current_suffix = 0 %} {% for seed in seeds.results %} {% set ns.seed_tokens = seed.msg.split(':') %} {% if ns.seed_tokens[2] != ns.current_rack %} {% set ns.current_rack = ns.seed_tokens[2] %} {% set ns.current_suffix = ns.current_suffix + 1 %} {% endif %} {{ ns.seed_tokens[0] }}:8102:{{ ns.seed_tokens[3] }}-{{ ns.current_suffix }}{%- if not loop.last %};{%- endif %} {%- endfor %} # Dynomite cluster name workflow.dynomite.cluster.name=dyno1 # Namespace for the keys stored in Dynomite/Redis workflow.namespace.prefix=conductor # Namespace prefix for the dyno queues workflow.namespace.queue.prefix=conductor_queues # No. of threads allocated to dyno-queues (optional) queues.dynomite.threads=10 # Non-quorum port used to connect to local redis. Used by dyno-queues. # When using redis directly, set this to the same port as redis server # For Dynomite, this is 22122 by default or the local redis-server port used by Dynomite. queues.dynomite.nonQuorum.port=22122 # Elastic search instance type. Possible values are memory and external. # If not specified, the instance type will be embedded in memory # # memory: The instance is created in memory and lost when the server dies. Useful for development and testing. # external: Elastic search instance runs outside of the server. Data is persisted and does not get lost when # the server dies. Useful for more stable environments like staging or production. workflow.elasticsearch.instanceType=external # Transport address to elasticsearch workflow.elasticsearch.url=elasticsearch:9300 # Name of the elasticsearch cluster workflow.elasticsearch.index.name=conductor # Additional modules (optional) # conductor.additional.modules=class_extending_com.google.inject.AbstractModule # Additional modules for metrics collection (optional) # conductor.additional.modules=com.netflix.conductor.contribs.metrics.MetricsRegistryModule,com.netflix.conductor.contribs.metrics.LoggingMetricsModule # com.netflix.conductor.contribs.metrics.LoggingMetricsModule.reportPeriodSeconds=15 # Load sample kitchen sink workflow loadSample=false