Documentation ¶
Overview ¶
Package dynamicconfig is a generated GoMock package.
Index ¶
- Constants
- Variables
- func GetAllKeys() map[string]Key
- func GetBoolPropertyFilteredByTaskListInfo(value bool) func(domain string, taskList string, taskType int) bool
- func GetBoolPropertyFn(value bool) func(opts ...FilterOption) bool
- func GetBoolPropertyFnFilteredByDomain(value bool) func(domain string) bool
- func GetBoolPropertyFnFilteredByDomainID(value bool) func(domainID string) bool
- func GetDurationPropertyFilteredByWorkflowType(value time.Duration) func(domainName string, workflowType string) time.Duration
- func GetDurationPropertyFn(value time.Duration) func(opts ...FilterOption) time.Duration
- func GetDurationPropertyFnFilteredByDomain(value time.Duration) func(domain string) time.Duration
- func GetDurationPropertyFnFilteredByShardID(value time.Duration) func(shardID int) time.Duration
- func GetDurationPropertyFnFilteredByTaskListInfo(value time.Duration) func(domain string, taskList string, taskType int) time.Duration
- func GetFloatPropertyFn(value float64) func(opts ...FilterOption) float64
- func GetIntPropertyFilteredByDomain(value int) func(domain string) int
- func GetIntPropertyFilteredByShardID(value int) func(shardID int) int
- func GetIntPropertyFilteredByTaskListInfo(value int) func(domain string, taskList string, taskType int) int
- func GetIntPropertyFilteredByWorkflowType(value int) func(domainName string, workflowType string) int
- func GetIntPropertyFn(value int) func(opts ...FilterOption) int
- func GetMapPropertyFn(value map[string]interface{}) func(opts ...FilterOption) map[string]interface{}
- func GetStringPropertyFn(value string) func(opts ...FilterOption) string
- func ToGetDynamicConfigFilterRequest(configName string, filters []FilterOption) *types.GetDynamicConfigRequest
- func ValidateKeyValuePair(key Key, value interface{}) error
- type BoolKey
- type BoolPropertyFn
- type BoolPropertyFnWithDomainFilter
- type BoolPropertyFnWithDomainIDAndWorkflowIDFilter
- type BoolPropertyFnWithDomainIDFilter
- type BoolPropertyFnWithTaskListInfoFilters
- type Client
- type Collection
- func (c *Collection) GetBoolProperty(key BoolKey) BoolPropertyFn
- func (c *Collection) GetBoolPropertyFilteredByDomain(key BoolKey) BoolPropertyFnWithDomainFilter
- func (c *Collection) GetBoolPropertyFilteredByDomainID(key BoolKey) BoolPropertyFnWithDomainIDFilter
- func (c *Collection) GetBoolPropertyFilteredByDomainIDAndWorkflowID(key BoolKey) BoolPropertyFnWithDomainIDAndWorkflowIDFilter
- func (c *Collection) GetBoolPropertyFilteredByTaskListInfo(key BoolKey) BoolPropertyFnWithTaskListInfoFilters
- func (c *Collection) GetDurationProperty(key DurationKey) DurationPropertyFn
- func (c *Collection) GetDurationPropertyFilteredByDomain(key DurationKey) DurationPropertyFnWithDomainFilter
- func (c *Collection) GetDurationPropertyFilteredByDomainID(key DurationKey) DurationPropertyFnWithDomainIDFilter
- func (c *Collection) GetDurationPropertyFilteredByShardID(key DurationKey) DurationPropertyFnWithShardIDFilter
- func (c *Collection) GetDurationPropertyFilteredByTaskListInfo(key DurationKey) DurationPropertyFnWithTaskListInfoFilters
- func (c *Collection) GetDurationPropertyFilteredByWorkflowType(key DurationKey) DurationPropertyFnWithWorkflowTypeFilter
- func (c *Collection) GetFloat64Property(key FloatKey) FloatPropertyFn
- func (c *Collection) GetFloat64PropertyFilteredByShardID(key FloatKey) FloatPropertyFnWithShardIDFilter
- func (c *Collection) GetIntProperty(key IntKey) IntPropertyFn
- func (c *Collection) GetIntPropertyFilteredByDomain(key IntKey) IntPropertyFnWithDomainFilter
- func (c *Collection) GetIntPropertyFilteredByShardID(key IntKey) IntPropertyFnWithShardIDFilter
- func (c *Collection) GetIntPropertyFilteredByTaskListInfo(key IntKey) IntPropertyFnWithTaskListInfoFilters
- func (c *Collection) GetIntPropertyFilteredByWorkflowType(key IntKey) IntPropertyFnWithWorkflowTypeFilter
- func (c *Collection) GetListProperty(key ListKey) ListPropertyFn
- func (c *Collection) GetMapProperty(key MapKey) MapPropertyFn
- func (c *Collection) GetProperty(key Key) PropertyFn
- func (c *Collection) GetStringProperty(key StringKey) StringPropertyFn
- func (c *Collection) GetStringPropertyFilteredByDomain(key StringKey) StringPropertyFnWithDomainFilter
- func (c *Collection) GetStringPropertyFilteredByRatelimitKey(key StringKey) StringPropertyWithRatelimitKeyFilter
- func (c *Collection) GetStringPropertyFilteredByTaskListInfo(key StringKey) StringPropertyFnWithTaskListInfoFilters
- type DurationKey
- type DurationPropertyFn
- type DurationPropertyFnWithDomainFilter
- type DurationPropertyFnWithDomainIDFilter
- type DurationPropertyFnWithShardIDFilter
- type DurationPropertyFnWithTaskListInfoFilters
- type DurationPropertyFnWithWorkflowTypeFilter
- type DynamicBool
- type DynamicDuration
- type DynamicFloat
- type DynamicInt
- type DynamicList
- type DynamicMap
- type DynamicString
- type FileBasedClientConfig
- type Filter
- type FilterOption
- func ClusterNameFilter(clusterName string) FilterOption
- func DomainFilter(name string) FilterOption
- func DomainIDFilter(domainID string) FilterOption
- func RatelimitKeyFilter(key string) FilterOption
- func ShardIDFilter(shardID int) FilterOption
- func TaskListFilter(name string) FilterOption
- func TaskTypeFilter(taskType int) FilterOption
- func WorkflowIDFilter(workflowID string) FilterOption
- func WorkflowTypeFilter(name string) FilterOption
- type FloatKey
- type FloatPropertyFn
- type FloatPropertyFnWithShardIDFilter
- type IntKey
- type IntPropertyFn
- type IntPropertyFnWithDomainFilter
- type IntPropertyFnWithShardIDFilter
- type IntPropertyFnWithTaskListInfoFilters
- type IntPropertyFnWithWorkflowTypeFilter
- type Key
- type ListKey
- type ListPropertyFn
- type MapKey
- type MapPropertyFn
- type MockClient
- func (m *MockClient) EXPECT() *MockClientMockRecorder
- func (m *MockClient) GetBoolValue(name BoolKey, filters map[Filter]interface{}) (bool, error)
- func (m *MockClient) GetDurationValue(name DurationKey, filters map[Filter]interface{}) (time.Duration, error)
- func (m *MockClient) GetFloatValue(name FloatKey, filters map[Filter]interface{}) (float64, error)
- func (m *MockClient) GetIntValue(name IntKey, filters map[Filter]interface{}) (int, error)
- func (m *MockClient) GetListValue(name ListKey, filters map[Filter]interface{}) ([]interface{}, error)
- func (m *MockClient) GetMapValue(name MapKey, filters map[Filter]interface{}) (map[string]interface{}, error)
- func (m *MockClient) GetStringValue(name StringKey, filters map[Filter]interface{}) (string, error)
- func (m *MockClient) GetValue(name Key) (interface{}, error)
- func (m *MockClient) GetValueWithFilters(name Key, filters map[Filter]interface{}) (interface{}, error)
- func (m *MockClient) ListValue(name Key) ([]*types.DynamicConfigEntry, error)
- func (m *MockClient) RestoreValue(name Key, filters map[Filter]interface{}) error
- func (m *MockClient) UpdateValue(name Key, value interface{}) error
- type MockClientMockRecorder
- func (mr *MockClientMockRecorder) GetBoolValue(name, filters interface{}) *gomock.Call
- func (mr *MockClientMockRecorder) GetDurationValue(name, filters interface{}) *gomock.Call
- func (mr *MockClientMockRecorder) GetFloatValue(name, filters interface{}) *gomock.Call
- func (mr *MockClientMockRecorder) GetIntValue(name, filters interface{}) *gomock.Call
- func (mr *MockClientMockRecorder) GetListValue(name, filters interface{}) *gomock.Call
- func (mr *MockClientMockRecorder) GetMapValue(name, filters interface{}) *gomock.Call
- func (mr *MockClientMockRecorder) GetStringValue(name, filters interface{}) *gomock.Call
- func (mr *MockClientMockRecorder) GetValue(name interface{}) *gomock.Call
- func (mr *MockClientMockRecorder) GetValueWithFilters(name, filters interface{}) *gomock.Call
- func (mr *MockClientMockRecorder) ListValue(name interface{}) *gomock.Call
- func (mr *MockClientMockRecorder) RestoreValue(name, filters interface{}) *gomock.Call
- func (mr *MockClientMockRecorder) UpdateValue(name, value interface{}) *gomock.Call
- type PropertyFn
- type StringKey
- type StringPropertyFn
- type StringPropertyFnWithDomainFilter
- type StringPropertyFnWithTaskListInfoFilters
- type StringPropertyWithRatelimitKeyFilter
Constants ¶
const ( ConfigStoreClient = "configstore" FileBasedClient = "filebased" InMemoryClient = "memory" NopClient = "nop" )
const UnlimitedRPS = math.MaxInt32
UnlimitedRPS represents an integer to use for "unlimited" RPS values.
Since our ratelimiters do int/float conversions, and zero or negative values result in not allowing any requests, math.MaxInt is unsafe:
int(float64(math.MaxInt)) // -9223372036854775808
Much higher values are possible, but we can't handle 2 billion RPS, this is good enough.
Variables ¶
var BoolKeys = map[BoolKey]DynamicBool{ TestGetBoolPropertyKey: { KeyName: "testGetBoolPropertyKey", Description: "", DefaultValue: false, }, TestGetBoolPropertyFilteredByDomainIDKey: { KeyName: "testGetBoolPropertyFilteredByDomainIDKey", Description: "", DefaultValue: false, }, TestGetBoolPropertyFilteredByTaskListInfoKey: { KeyName: "testGetBoolPropertyFilteredByTaskListInfoKey", Description: "", DefaultValue: false, }, TestGetBoolPropertyFilteredByDomainKey: { KeyName: "testGetBoolPropertyFilteredByDomainKey", Description: "", DefaultValue: false, Filters: nil, }, TestGetBoolPropertyFilteredByDomainIDAndWorkflowIDKey: { KeyName: "testGetBoolPropertyFilteredByDomainIDandWorkflowIDKey", Description: "", DefaultValue: false, Filters: nil, }, EnableVisibilitySampling: { KeyName: "system.enableVisibilitySampling", Description: "EnableVisibilitySampling is key for enable visibility sampling for basic(DB based) visibility", DefaultValue: false, Filters: nil, }, EnableReadFromClosedExecutionV2: { KeyName: "system.enableReadFromClosedExecutionV2", Description: "EnableReadFromClosedExecutionV2 is key for enable read from cadence_visibility.closed_executions_v2", DefaultValue: false, }, EnableReadVisibilityFromES: { KeyName: "system.enableReadVisibilityFromES", Filters: []Filter{DomainName}, Description: "EnableReadVisibilityFromES is key for enable read from elastic search or db visibility, usually using with AdvancedVisibilityWritingMode for seamless migration from db visibility to advanced visibility", DefaultValue: true, }, EnableReadVisibilityFromPinot: { KeyName: "system.enableReadVisibilityFromPinot", Filters: []Filter{DomainName}, Description: "EnableReadVisibilityFromPinot is key for enable read from pinot or db visibility, usually using with AdvancedVisibilityWritingMode for seamless migration from db visibility to advanced visibility", DefaultValue: true, }, EnableVisibilityDoubleRead: { KeyName: "system.enableVisibilityDoubleRead", Filters: []Filter{DomainName}, Description: "EnableVisibilityDoubleRead is key for enable read for both elastic search and Pinot for a latency comparison", DefaultValue: false, }, EnableLogCustomerQueryParameter: { KeyName: "system.enableLogCustomerQueryParameter", Filters: []Filter{DomainName}, Description: "EnableLogCustomerQueryParameter is key for enable log customer query parameters", DefaultValue: false, }, EmitShardDiffLog: { KeyName: "history.emitShardDiffLog", Description: "EmitShardDiffLog is whether emit the shard diff log", DefaultValue: false, }, EnableRecordWorkflowExecutionUninitialized: { KeyName: "history.enableRecordWorkflowExecutionUninitialized", Description: "EnableRecordWorkflowExecutionUninitialized enables record workflow execution uninitialized state in ElasticSearch", DefaultValue: false, }, DisableListVisibilityByFilter: { KeyName: "frontend.disableListVisibilityByFilter", Filters: []Filter{DomainName}, Description: "DisableListVisibilityByFilter is config to disable list open/close workflow using filter", DefaultValue: false, }, EnableReadFromHistoryArchival: { KeyName: "system.enableReadFromHistoryArchival", Description: "EnableReadFromHistoryArchival is key for enabling reading history from archival store", DefaultValue: true, }, EnableReadFromVisibilityArchival: { KeyName: "system.enableReadFromVisibilityArchival", Description: "EnableReadFromVisibilityArchival is key for enabling reading visibility from archival store to override the value from static config.", DefaultValue: true, }, EnableDomainNotActiveAutoForwarding: { KeyName: "system.enableDomainNotActiveAutoForwarding", Filters: []Filter{DomainName}, Description: "EnableDomainNotActiveAutoForwarding decides requests form which domain will be forwarded to active cluster if domain is not active in current cluster. Only when selected-api-forwarding or all-domain-apis-forwarding is the policy in ClusterRedirectionPolicy(in static config). If the policy is noop(default) this flag is not doing anything.", DefaultValue: true, }, EnableGracefulFailover: { KeyName: "system.enableGracefulFailover", Description: "EnableGracefulFailover is whether enabling graceful failover", DefaultValue: true, }, DisallowQuery: { KeyName: "system.disallowQuery", Filters: []Filter{DomainName}, Description: "DisallowQuery is the key to disallow query for a domain", DefaultValue: false, }, EnableDebugMode: { KeyName: "system.enableDebugMode", Description: "EnableDebugMode is for enabling debugging components, logs and metrics", DefaultValue: false, }, EnableGRPCOutbound: { KeyName: "system.enableGRPCOutbound", Description: "EnableGRPCOutbound is the key for enabling outbound GRPC traffic", DefaultValue: true, }, EnableSQLAsyncTransaction: { KeyName: "system.enableSQLAsyncTransaction", Description: "EnableSQLAsyncTransaction is the key for enabling async transaction", DefaultValue: false, }, EnableConnectionRetainingDirectChooser: { KeyName: "system.enableConnectionRetainingDirectChooser", Description: "EnableConnectionRetainingDirectChooser is the key for enabling connection retaining direct chooser", DefaultValue: false, }, EnableClientVersionCheck: { KeyName: "frontend.enableClientVersionCheck", Description: "EnableClientVersionCheck is enables client version check for frontend", DefaultValue: false, }, SendRawWorkflowHistory: { KeyName: "frontend.sendRawWorkflowHistory", Filters: []Filter{DomainName}, Description: "SendRawWorkflowHistory is whether to enable raw history retrieving", DefaultValue: false, }, FrontendEmitSignalNameMetricsTag: { KeyName: "frontend.emitSignalNameMetricsTag", Filters: []Filter{DomainName}, Description: "FrontendEmitSignalNameMetricsTag enables emitting signal name tag in metrics in frontend client", DefaultValue: false, }, EnableQueryAttributeValidation: { KeyName: "frontend.enableQueryAttributeValidation", Description: "EnableQueryAttributeValidation enables validation of queries' search attributes against the dynamic config whitelist", DefaultValue: true, }, MatchingEnableSyncMatch: { KeyName: "matching.enableSyncMatch", Filters: []Filter{DomainName, TaskListName, TaskType}, Description: "MatchingEnableSyncMatch is to enable sync match", DefaultValue: true, }, MatchingEnableTaskInfoLogByDomainID: { KeyName: "matching.enableTaskInfoLogByDomainID", Filters: []Filter{DomainID}, Description: "MatchingEnableTaskInfoLogByDomainID is enables info level logs for decision/activity task based on the request domainID", DefaultValue: false, }, MatchingEnableTasklistGuardAgainstOwnershipShardLoss: { KeyName: "matching.enableTasklistGuardAgainstOwnershipLoss", Description: "allows guards to ensure that tasklists don't continue processing if there's signal that they've lost ownership", DefaultValue: false, }, MatchingEnableGetNumberOfPartitionsFromCache: { KeyName: "matching.enableGetNumberOfPartitionsFromCache", Filters: []Filter{DomainName, TaskListName, TaskType}, Description: "MatchingEnableGetNumberOfPartitionsFromCache is to enable getting number of partitions from cache instead of dynamic config", DefaultValue: false, }, EventsCacheGlobalEnable: { KeyName: "history.eventsCacheGlobalEnable", Description: "EventsCacheGlobalEnable is enables global cache over all history shards", DefaultValue: false, }, QueueProcessorEnableSplit: { KeyName: "history.queueProcessorEnableSplit", Description: "QueueProcessorEnableSplit indicates whether processing queue split policy should be enabled", DefaultValue: false, }, QueueProcessorEnableRandomSplitByDomainID: { KeyName: "history.queueProcessorEnableRandomSplitByDomainID", Filters: []Filter{DomainID}, Description: "QueueProcessorEnableRandomSplitByDomainID indicates whether random queue split policy should be enabled for a domain", DefaultValue: false, }, QueueProcessorEnablePendingTaskSplitByDomainID: { KeyName: "history.queueProcessorEnablePendingTaskSplitByDomainID", Filters: []Filter{DomainID}, Description: "ueueProcessorEnablePendingTaskSplitByDomainID indicates whether pending task split policy should be enabled", DefaultValue: false, }, QueueProcessorEnableStuckTaskSplitByDomainID: { KeyName: "history.queueProcessorEnableStuckTaskSplitByDomainID", Filters: []Filter{DomainID}, Description: "QueueProcessorEnableStuckTaskSplitByDomainID indicates whether stuck task split policy should be enabled", DefaultValue: false, }, QueueProcessorEnablePersistQueueStates: { KeyName: "history.queueProcessorEnablePersistQueueStates", Description: "QueueProcessorEnablePersistQueueStates indicates whether processing queue states should be persisted", DefaultValue: true, }, QueueProcessorEnableLoadQueueStates: { KeyName: "history.queueProcessorEnableLoadQueueStates", Description: "QueueProcessorEnableLoadQueueStates indicates whether processing queue states should be loaded", DefaultValue: true, }, QueueProcessorEnableGracefulSyncShutdown: { KeyName: "history.queueProcessorEnableGracefulSyncShutdown", Description: "QueueProcessorEnableGracefulSyncShutdown indicates whether processing queue should be shutdown gracefully & synchronously", DefaultValue: false, }, ReplicationTaskFetcherEnableGracefulSyncShutdown: { KeyName: "history.replicationTaskFetcherEnableGracefulSyncShutdown", Description: "ReplicationTaskFetcherEnableGracefulSyncShutdown is whether we should gracefully drain replication task fetcher on shutdown", DefaultValue: false, }, TransferProcessorEnableValidator: { KeyName: "history.transferProcessorEnableValidator", Description: "TransferProcessorEnableValidator is whether validator should be enabled for transferQueueProcessor", DefaultValue: false, }, EnableAdminProtection: { KeyName: "history.enableAdminProtection", Description: "EnableAdminProtection is whether to enable admin checking", DefaultValue: false, }, EnableParentClosePolicy: { KeyName: "history.enableParentClosePolicy", Filters: []Filter{DomainName}, Description: "EnableParentClosePolicy is whether to ParentClosePolicy", DefaultValue: true, }, EnableDropStuckTaskByDomainID: { KeyName: "history.DropStuckTaskByDomain", Filters: []Filter{DomainID}, Description: "EnableDropStuckTaskByDomainID is whether stuck timer/transfer task should be dropped for a domain", DefaultValue: false, }, EnableConsistentQuery: { KeyName: "history.EnableConsistentQuery", Description: "EnableConsistentQuery indicates if consistent query is enabled for the cluster", DefaultValue: true, }, EnableConsistentQueryByDomain: { KeyName: "history.EnableConsistentQueryByDomain", Filters: []Filter{DomainName}, Description: "EnableConsistentQueryByDomain indicates if consistent query is enabled for a domain", DefaultValue: false, }, EnableContextHeaderInVisibility: { KeyName: "history.enableContextHeaderInVisibility", Filters: []Filter{DomainName}, Description: "EnableContextHeaderInVisibility is key for enable context header in visibility", DefaultValue: false, }, EnableCrossClusterOperationsForDomain: { KeyName: "history.enableCrossClusterOperations", Filters: []Filter{DomainName}, Description: "EnableCrossClusterOperationsForDomain indicates if cross cluster operations can be scheduled for a domain", DefaultValue: false, }, EnableHistoryCorruptionCheck: { KeyName: "history.enableHistoryCorruptionCheck", Filters: []Filter{DomainName}, Description: "EnableHistoryCorruptionCheck enables additional sanity check for corrupted history. This allows early catches of DB corruptions but potiantally increased latency.", DefaultValue: false, }, EnableActivityLocalDispatchByDomain: { KeyName: "history.enableActivityLocalDispatchByDomain", Filters: []Filter{DomainName}, Description: "EnableActivityLocalDispatchByDomain is allows worker to dispatch activity tasks through local tunnel after decisions are made. This is an performance optimization to skip activity scheduling efforts", DefaultValue: true, }, HistoryEnableTaskInfoLogByDomainID: { KeyName: "history.enableTaskInfoLogByDomainID", Filters: []Filter{DomainID}, Description: "HistoryEnableTaskInfoLogByDomainID is enables info level logs for decision/activity task based on the request domainID", DefaultValue: false, }, EnableReplicationTaskGeneration: { KeyName: "history.enableReplicationTaskGeneration", Filters: []Filter{DomainID, WorkflowID}, Description: "EnableReplicationTaskGeneration is the flag to control replication generation", DefaultValue: true, }, UseNewInitialFailoverVersion: { KeyName: "history.useNewInitialFailoverVersion", Description: "use the minInitialFailover version", DefaultValue: false, }, AllowArchivingIncompleteHistory: { KeyName: "worker.AllowArchivingIncompleteHistory", Description: "AllowArchivingIncompleteHistory will continue on when seeing some error like history mutated(usually caused by database consistency issues)", DefaultValue: false, }, EnableCleaningOrphanTaskInTasklistScavenger: { KeyName: "worker.enableCleaningOrphanTaskInTasklistScavenger", Description: "EnableCleaningOrphanTaskInTasklistScavenger indicates if enabling the scanner to clean up orphan tasks", DefaultValue: false, }, TaskListScannerEnabled: { KeyName: "worker.taskListScannerEnabled", Description: "TaskListScannerEnabled indicates if task list scanner should be started as part of worker.Scanner", DefaultValue: true, }, HistoryScannerEnabled: { KeyName: "worker.historyScannerEnabled", Description: "HistoryScannerEnabled indicates if history scanner should be started as part of worker.Scanner", DefaultValue: false, }, ConcreteExecutionsScannerEnabled: { KeyName: "worker.executionsScannerEnabled", Description: "ConcreteExecutionsScannerEnabled indicates if executions scanner should be started as part of worker.Scanner", DefaultValue: false, }, ConcreteExecutionsScannerInvariantCollectionMutableState: { KeyName: "worker.executionsScannerInvariantCollectionMutableState", Description: "ConcreteExecutionsScannerInvariantCollectionMutableState indicates if mutable state invariant checks should be run", DefaultValue: true, }, ConcreteExecutionsScannerInvariantCollectionHistory: { KeyName: "worker.executionsScannerInvariantCollectionHistory", Description: "ConcreteExecutionsScannerInvariantCollectionHistory indicates if history invariant checks should be run", DefaultValue: true, }, ConcreteExecutionsScannerInvariantCollectionStale: { KeyName: "worker.executionsScannerInvariantCollectionStale", Description: "ConcreteExecutionsScannerInvariantCollectionStale indicates if the stale-workflow invariant should be run", DefaultValue: false, }, ConcreteExecutionsFixerInvariantCollectionMutableState: { KeyName: "worker.executionsFixerInvariantCollectionMutableState", Description: "ConcreteExecutionsFixerInvariantCollectionMutableState indicates if mutable state invariant checks should be run", DefaultValue: true, }, ConcreteExecutionsFixerInvariantCollectionHistory: { KeyName: "worker.executionsFixerInvariantCollectionHistory", Description: "ConcreteExecutionsFixerInvariantCollectionHistory indicates if history invariant checks should be run", DefaultValue: true, }, ConcreteExecutionsFixerInvariantCollectionStale: { KeyName: "worker.executionsFixerInvariantCollectionStale", Description: "ConcreteExecutionsFixerInvariantCollectionStale indicates if the stale-workflow invariant should be run", DefaultValue: false, }, CurrentExecutionsScannerEnabled: { KeyName: "worker.currentExecutionsScannerEnabled", Description: "CurrentExecutionsScannerEnabled indicates if current executions scanner should be started as part of worker.Scanner", DefaultValue: false, }, CurrentExecutionsScannerInvariantCollectionHistory: { KeyName: "worker.currentExecutionsScannerInvariantCollectionHistory", Description: "CurrentExecutionsScannerInvariantCollectionHistory indicates if history invariant checks should be run", DefaultValue: true, }, CurrentExecutionsScannerInvariantCollectionMutableState: { KeyName: "worker.currentExecutionsInvariantCollectionMutableState", Description: "CurrentExecutionsScannerInvariantCollectionMutableState indicates if mutable state invariant checks should be run", DefaultValue: true, }, EnableBatcher: { KeyName: "worker.enableBatcher", Description: "EnableBatcher is decides whether start batcher in our worker", DefaultValue: true, }, EnableParentClosePolicyWorker: { KeyName: "system.enableParentClosePolicyWorker", Description: "EnableParentClosePolicyWorker decides whether or not enable system workers for processing parent close policy task", DefaultValue: true, }, EnableESAnalyzer: { KeyName: "system.enableESAnalyzer", Description: "EnableESAnalyzer decides whether to enable system workers for processing ElasticSearch Analyzer", DefaultValue: false, }, EnableAsyncWorkflowConsumption: { KeyName: "worker.enableAsyncWorkflowConsumption", Description: "EnableAsyncWorkflowConsumption decides whether to enable async workflows", DefaultValue: false, }, EnableStickyQuery: { KeyName: "system.enableStickyQuery", Filters: []Filter{DomainName}, Description: "EnableStickyQuery indicates if sticky query should be enabled per domain", DefaultValue: true, }, EnableFailoverManager: { KeyName: "system.enableFailoverManager", Description: "EnableFailoverManager indicates if failover manager is enabled", DefaultValue: true, }, ConcreteExecutionFixerDomainAllow: { KeyName: "worker.concreteExecutionFixerDomainAllow", Filters: []Filter{DomainName}, Description: "ConcreteExecutionFixerDomainAllow is which domains are allowed to be fixed by concrete fixer workflow", DefaultValue: false, }, CurrentExecutionFixerDomainAllow: { KeyName: "worker.currentExecutionFixerDomainAllow", Filters: []Filter{DomainName}, Description: "CurrentExecutionFixerDomainAllow is which domains are allowed to be fixed by current fixer workflow", DefaultValue: false, }, TimersScannerEnabled: { KeyName: "worker.timersScannerEnabled", Description: "TimersScannerEnabled is if timers scanner should be started as part of worker.Scanner", DefaultValue: false, }, TimersFixerEnabled: { KeyName: "worker.timersFixerEnabled", Description: "TimersFixerEnabled is if timers fixer should be started as part of worker.Scanner", DefaultValue: false, }, TimersFixerDomainAllow: { KeyName: "worker.timersFixerDomainAllow", Filters: []Filter{DomainName}, Description: "TimersFixerDomainAllow is which domains are allowed to be fixed by timer fixer workflow", DefaultValue: false, }, ConcreteExecutionFixerEnabled: { KeyName: "worker.concreteExecutionFixerEnabled", Description: "ConcreteExecutionFixerEnabled is if concrete execution fixer workflow is enabled", DefaultValue: false, }, CurrentExecutionFixerEnabled: { KeyName: "worker.currentExecutionFixerEnabled", Description: "CurrentExecutionFixerEnabled is if current execution fixer workflow is enabled", DefaultValue: false, }, EnableAuthorization: { KeyName: "system.enableAuthorization", Description: "EnableAuthorization is the key to enable authorization for a domain, only for extension binary:", DefaultValue: false, }, EnableTasklistIsolation: { KeyName: "system.enableTasklistIsolation", Description: "EnableTasklistIsolation is a feature to enable isolation-groups for a domain. Should not be enabled without a deep understanding of this feature", DefaultValue: false, }, EnableServiceAuthorization: { KeyName: "system.enableServiceAuthorization", Description: "EnableServiceAuthorization is the key to enable authorization for a service, only for extension binary:", DefaultValue: false, }, EnableServiceAuthorizationLogOnly: { KeyName: "system.enableServiceAuthorizationLogOnly", Description: "EnableServiceAuthorizationLogOnly is the key to enable authorization logging for a service, only for extension binary:", DefaultValue: false, }, ESAnalyzerPause: { KeyName: "worker.ESAnalyzerPause", Description: "ESAnalyzerPause defines if we want to dynamically pause the analyzer workflow", DefaultValue: false, }, EnableArchivalCompression: { KeyName: "worker.EnableArchivalCompression", Description: "EnableArchivalCompression indicates whether blobs are compressed before they are archived", DefaultValue: false, }, ESAnalyzerEnableAvgDurationBasedChecks: { KeyName: "worker.ESAnalyzerEnableAvgDurationBasedChecks", Description: "ESAnalyzerEnableAvgDurationBasedChecks controls if we want to enable avg duration based task refreshes", DefaultValue: false, }, Lockdown: { KeyName: "system.Lockdown", Description: "Lockdown defines if we want to allow failovers of domains to this cluster", DefaultValue: false, }, EnablePendingActivityValidation: { KeyName: "limit.pendingActivityCount.enabled", Description: "Enables pending activity count limiting/validation", DefaultValue: false, }, EnableCassandraAllConsistencyLevelDelete: { KeyName: "system.enableCassandraAllConsistencyLevelDelete", Description: "Uses all consistency level for Cassandra delete operations", DefaultValue: false, }, EnableShardIDMetrics: { KeyName: "system.enableShardIDMetrics", Description: "Enable shardId metrics in persistence client", DefaultValue: true, }, EnableTimerDebugLogByDomainID: { KeyName: "history.enableTimerDebugLogByDomainID", Filters: []Filter{DomainID}, Description: "Enable log for debugging timer task issue by domain", DefaultValue: false, }, EnableTaskVal: { KeyName: "system.enableTaskVal", Description: "Enable TaskValidation", DefaultValue: false, }, WorkflowIDCacheExternalEnabled: { KeyName: "history.workflowIDCacheExternalEnabled", Filters: []Filter{DomainName}, Description: "WorkflowIDCacheExternalEnabled is the key to enable/disable caching of workflowID specific information for external requests", DefaultValue: false, }, WorkflowIDCacheInternalEnabled: { KeyName: "history.workflowIDCacheInternalEnabled", Filters: []Filter{DomainName}, Description: "WorkflowIDCacheInternalEnabled is the key to enable/disable caching of workflowID specific information for internal requests", DefaultValue: false, }, WorkflowIDExternalRateLimitEnabled: { KeyName: "history.workflowIDExternalRateLimitEnabled", Filters: []Filter{DomainName}, Description: "WorkflowIDExternalRateLimitEnabled is the key to enable/disable rate limiting of specific workflowIDs for external requests", DefaultValue: false, }, WorkflowIDInternalRateLimitEnabled: { KeyName: "history.workflowIDInternalRateLimitEnabled", Filters: []Filter{DomainName}, Description: "WorkflowIDInternalRateLimitEnabled is the key to enable/disable rate limiting of specific workflowIDs for internal requests", DefaultValue: false, }, EnableRetryForChecksumFailure: { KeyName: "history.enableMutableStateChecksumFailureRetry", Filters: []Filter{DomainName}, Description: "EnableRetryForChecksumFailure enables retry if mutable state checksum verification fails", DefaultValue: false, }, EnableStrongIdempotency: DynamicBool{ KeyName: "history.enableStrongIdempotency", Filters: []Filter{DomainName}, Description: "EnableStrongIdempotency enables strong idempotency for APIs", DefaultValue: false, }, EnableStrongIdempotencySanityCheck: DynamicBool{ KeyName: "history.enableStrongIdempotencySanityCheck", Filters: []Filter{DomainName}, Description: "EnableStrongIdempotencySanityCheck enables sanity check for strong idempotency", DefaultValue: false, }, }
var DurationKeys = map[DurationKey]DynamicDuration{ TestGetDurationPropertyKey: { KeyName: "testGetDurationPropertyKey", Description: "", DefaultValue: 0, }, TestGetDurationPropertyFilteredByDomainKey: { KeyName: "testGetDurationPropertyFilteredByDomainKey", Description: "", DefaultValue: 0, }, TestGetDurationPropertyFilteredByTaskListInfoKey: { KeyName: "testGetDurationPropertyFilteredByTaskListInfoKey", Description: "", DefaultValue: 0, }, TestGetDurationPropertyFilteredByWorkflowTypeKey: { KeyName: "testGetDurationPropertyFilteredByWorkflowTypeKey", Description: "", DefaultValue: 0, Filters: nil, }, TestGetDurationPropertyFilteredByDomainIDKey: { KeyName: "testGetDurationPropertyFilteredByDomainIDKey", Description: "", DefaultValue: 0, Filters: nil, }, TestGetDurationPropertyFilteredByShardID: { KeyName: "testGetDurationPropertyFilteredByShardID", Description: "", DefaultValue: 0, Filters: nil, }, FrontendShutdownDrainDuration: { KeyName: "frontend.shutdownDrainDuration", Description: "FrontendShutdownDrainDuration is the duration of traffic drain during shutdown", DefaultValue: 0, }, FrontendFailoverCoolDown: { KeyName: "frontend.failoverCoolDown", Filters: []Filter{DomainName}, Description: "FrontendFailoverCoolDown is duration between two domain failvoers", DefaultValue: time.Minute, }, DomainFailoverRefreshInterval: { KeyName: "frontend.domainFailoverRefreshInterval", Description: "DomainFailoverRefreshInterval is the domain failover refresh timer", DefaultValue: time.Second * 10, }, GlobalRatelimiterUpdateInterval: { KeyName: "frontend.globalRatelimiterUpdateInterval", Description: "GlobalRatelimiterUpdateInterval defines how often each global ratelimiter collection submits load information, and the expected update rate in aggregators (used to determine when hosts are lost)", DefaultValue: 3 * time.Second, }, MatchingLongPollExpirationInterval: { KeyName: "matching.longPollExpirationInterval", Filters: []Filter{DomainName, TaskListName, TaskType}, Description: "MatchingLongPollExpirationInterval is the long poll expiration interval in the matching service", DefaultValue: time.Minute, }, MatchingUpdateAckInterval: { KeyName: "matching.updateAckInterval", Filters: []Filter{DomainName, TaskListName, TaskType}, Description: "MatchingUpdateAckInterval is the interval for update ack", DefaultValue: time.Minute, }, MatchingIdleTasklistCheckInterval: { KeyName: "matching.idleTasklistCheckInterval", Filters: []Filter{DomainName, TaskListName, TaskType}, Description: "MatchingIdleTasklistCheckInterval is the IdleTasklistCheckInterval", DefaultValue: time.Minute * 5, }, MaxTasklistIdleTime: { KeyName: "matching.maxTasklistIdleTime", Filters: []Filter{DomainName, TaskListName, TaskType}, Description: "MaxTasklistIdleTime is the max time tasklist being idle", DefaultValue: time.Minute * 5, }, MatchingShutdownDrainDuration: { KeyName: "matching.shutdownDrainDuration", Description: "MatchingShutdownDrainDuration is the duration of traffic drain during shutdown", DefaultValue: 0, }, MatchingActivityTaskSyncMatchWaitTime: { KeyName: "matching.activityTaskSyncMatchWaitTime", Filters: []Filter{DomainName}, Description: "MatchingActivityTaskSyncMatchWaitTime is the amount of time activity task will wait to be sync matched", DefaultValue: time.Millisecond * 50, }, HistoryLongPollExpirationInterval: { KeyName: "history.longPollExpirationInterval", Filters: []Filter{DomainName}, Description: "HistoryLongPollExpirationInterval is the long poll expiration interval in the history service", DefaultValue: time.Second * 20, }, HistoryCacheTTL: { KeyName: "history.cacheTTL", Description: "HistoryCacheTTL is TTL of history cache", DefaultValue: time.Hour, }, HistoryShutdownDrainDuration: { KeyName: "history.shutdownDrainDuration", Description: "HistoryShutdownDrainDuration is the duration of traffic drain during shutdown", DefaultValue: 0, }, EventsCacheTTL: { KeyName: "history.eventsCacheTTL", Description: "EventsCacheTTL is TTL of events cache", DefaultValue: time.Hour, }, AcquireShardInterval: { KeyName: "history.acquireShardInterval", Description: "AcquireShardInterval is interval that timer used to acquire shard", DefaultValue: time.Minute, }, StandbyClusterDelay: { KeyName: "history.standbyClusterDelay", Description: "StandbyClusterDelay is the artificial delay added to standby cluster's view of active cluster's time", DefaultValue: time.Minute * 5, }, StandbyTaskMissingEventsResendDelay: { KeyName: "history.standbyTaskMissingEventsResendDelay", Description: "StandbyTaskMissingEventsResendDelay is the amount of time standby cluster's will wait (if events are missing)before calling remote for missing events", DefaultValue: time.Minute * 15, }, StandbyTaskMissingEventsDiscardDelay: { KeyName: "history.standbyTaskMissingEventsDiscardDelay", Description: "StandbyTaskMissingEventsDiscardDelay is the amount of time standby cluster's will wait (if events are missing)before discarding the task", DefaultValue: time.Minute * 25, }, ActiveTaskRedispatchInterval: { KeyName: "history.activeTaskRedispatchInterval", Description: "ActiveTaskRedispatchInterval is the active task redispatch interval", DefaultValue: time.Second * 5, }, StandbyTaskRedispatchInterval: { KeyName: "history.standbyTaskRedispatchInterval", Description: "StandbyTaskRedispatchInterval is the standby task redispatch interval", DefaultValue: time.Second * 30, }, StandbyTaskReReplicationContextTimeout: { KeyName: "history.standbyTaskReReplicationContextTimeout", Filters: []Filter{DomainID}, Description: "StandbyTaskReReplicationContextTimeout is the context timeout for standby task re-replication", DefaultValue: time.Minute * 3, }, ResurrectionCheckMinDelay: { KeyName: "history.resurrectionCheckMinDelay", Filters: []Filter{DomainName}, Description: "ResurrectionCheckMinDelay is the minimal timer processing delay before scanning history to see if there's a resurrected timer/activity", DefaultValue: time.Hour * 24, }, QueueProcessorSplitLookAheadDurationByDomainID: { KeyName: "history.queueProcessorSplitLookAheadDurationByDomainID", Filters: []Filter{DomainID}, Description: "QueueProcessorSplitLookAheadDurationByDomainID is the look ahead duration when spliting a domain to a new processing queue", DefaultValue: time.Minute * 20, }, QueueProcessorPollBackoffInterval: { KeyName: "history.queueProcessorPollBackoffInterval", Description: "QueueProcessorPollBackoffInterval is the backoff duration when queue processor is throttled", DefaultValue: time.Second * 5, }, TimerProcessorUpdateAckInterval: { KeyName: "history.timerProcessorUpdateAckInterval", Description: "TimerProcessorUpdateAckInterval is update interval for timer processor", DefaultValue: time.Second * 30, }, TimerProcessorCompleteTimerInterval: { KeyName: "history.timerProcessorCompleteTimerInterval", Description: "TimerProcessorCompleteTimerInterval is complete timer interval for timer processor", DefaultValue: time.Minute, }, TimerProcessorFailoverMaxStartJitterInterval: { KeyName: "history.timerProcessorFailoverMaxStartJitterInterval", Description: "TimerProcessorFailoverMaxStartJitterInterval is the max jitter interval for starting timer failover queue processing. The actual jitter interval used will be a random duration between 0 and the max interval so that timer failover queue across different shards won't start at the same time", DefaultValue: 0, }, TimerProcessorMaxPollInterval: { KeyName: "history.timerProcessorMaxPollInterval", Description: "TimerProcessorMaxPollInterval is max poll interval for timer processor", DefaultValue: time.Minute * 5, }, TimerProcessorSplitQueueInterval: { KeyName: "history.timerProcessorSplitQueueInterval", Description: "TimerProcessorSplitQueueInterval is the split processing queue interval for timer processor", DefaultValue: time.Minute, }, TimerProcessorArchivalTimeLimit: { KeyName: "history.timerProcessorArchivalTimeLimit", Description: "TimerProcessorArchivalTimeLimit is the upper time limit for inline history archival", DefaultValue: time.Second * 2, }, TimerProcessorMaxTimeShift: { KeyName: "history.timerProcessorMaxTimeShift", Description: "TimerProcessorMaxTimeShift is the max shift timer processor can have", DefaultValue: time.Second, }, TransferProcessorFailoverMaxStartJitterInterval: { KeyName: "history.transferProcessorFailoverMaxStartJitterInterval", Description: "TransferProcessorFailoverMaxStartJitterInterval is the max jitter interval for starting transfer failover queue processing. The actual jitter interval used will be a random duration between 0 and the max interval so that timer failover queue across different shards won't start at the same time", DefaultValue: 0, }, TransferProcessorMaxPollInterval: { KeyName: "history.transferProcessorMaxPollInterval", Description: "TransferProcessorMaxPollInterval is max poll interval for transferQueueProcessor", DefaultValue: time.Minute, }, TransferProcessorSplitQueueInterval: { KeyName: "history.transferProcessorSplitQueueInterval", Description: "TransferProcessorSplitQueueInterval is the split processing queue interval for transferQueueProcessor", DefaultValue: time.Minute, }, TransferProcessorUpdateAckInterval: { KeyName: "history.transferProcessorUpdateAckInterval", Description: "TransferProcessorUpdateAckInterval is update interval for transferQueueProcessor", DefaultValue: time.Second * 30, }, TransferProcessorCompleteTransferInterval: { KeyName: "history.transferProcessorCompleteTransferInterval", Description: "TransferProcessorCompleteTransferInterval is complete timer interval for transferQueueProcessor", DefaultValue: time.Minute, }, TransferProcessorValidationInterval: { KeyName: "history.transferProcessorValidationInterval", Description: "TransferProcessorValidationInterval is interval for performing transfer queue validation", DefaultValue: time.Second * 30, }, TransferProcessorVisibilityArchivalTimeLimit: { KeyName: "history.transferProcessorVisibilityArchivalTimeLimit", Description: "TransferProcessorVisibilityArchivalTimeLimit is the upper time limit for archiving visibility records", DefaultValue: time.Millisecond * 400, }, ReplicatorUpperLatency: { KeyName: "history.replicatorUpperLatency", Description: "ReplicatorUpperLatency indicates the max allowed replication latency between clusters", DefaultValue: time.Second * 40, }, ShardUpdateMinInterval: { KeyName: "history.shardUpdateMinInterval", Description: "ShardUpdateMinInterval is the minimal time interval which the shard info can be updated", DefaultValue: time.Minute * 5, }, ShardSyncMinInterval: { KeyName: "history.shardSyncMinInterval", Description: "ShardSyncMinInterval is the minimal time interval which the shard info should be sync to remote", DefaultValue: time.Minute * 5, }, StickyTTL: { KeyName: "history.stickyTTL", Filters: []Filter{DomainName}, Description: "StickyTTL is to expire a sticky tasklist if no update more than this duration", DefaultValue: time.Hour * 24 * 365, }, DecisionHeartbeatTimeout: { KeyName: "history.decisionHeartbeatTimeout", Filters: []Filter{DomainName}, Description: "DecisionHeartbeatTimeout is for decision heartbeat", DefaultValue: time.Minute * 30, }, NormalDecisionScheduleToStartTimeout: { KeyName: "history.normalDecisionScheduleToStartTimeout", Filters: []Filter{DomainName}, Description: "NormalDecisionScheduleToStartTimeout is scheduleToStart timeout duration for normal (non-sticky) decision task", DefaultValue: time.Minute * 5, }, NotifyFailoverMarkerInterval: { KeyName: "history.NotifyFailoverMarkerInterval", Description: "NotifyFailoverMarkerInterval is determines the frequency to notify failover marker", DefaultValue: time.Second * 5, }, ActivityMaxScheduleToStartTimeoutForRetry: { KeyName: "history.activityMaxScheduleToStartTimeoutForRetry", Filters: []Filter{DomainName}, Description: "ActivityMaxScheduleToStartTimeoutForRetry is maximum value allowed when overwritting the schedule to start timeout for activities with retry policy", DefaultValue: time.Minute * 30, }, ReplicationTaskFetcherAggregationInterval: { KeyName: "history.ReplicationTaskFetcherAggregationInterval", Description: "ReplicationTaskFetcherAggregationInterval determines how frequently the fetch requests are sent", DefaultValue: time.Second * 2, }, ReplicationTaskFetcherErrorRetryWait: { KeyName: "history.ReplicationTaskFetcherErrorRetryWait", Description: "ReplicationTaskFetcherErrorRetryWait is the wait time when fetcher encounters error", DefaultValue: time.Second, }, ReplicationTaskFetcherServiceBusyWait: { KeyName: "history.ReplicationTaskFetcherServiceBusyWait", Description: "ReplicationTaskFetcherServiceBusyWait is the wait time when fetcher encounters service busy error", DefaultValue: time.Minute, }, ReplicationTaskProcessorErrorRetryWait: { KeyName: "history.ReplicationTaskProcessorErrorRetryWait", Filters: []Filter{ShardID}, Description: "ReplicationTaskProcessorErrorRetryWait is the initial retry wait when we see errors in applying replication tasks", DefaultValue: time.Millisecond * 50, }, ReplicationTaskProcessorErrorSecondRetryWait: { KeyName: "history.ReplicationTaskProcessorErrorSecondRetryWait", Filters: []Filter{ShardID}, Description: "ReplicationTaskProcessorErrorSecondRetryWait is the initial retry wait for the second phase retry", DefaultValue: time.Second * 5, }, ReplicationTaskProcessorErrorSecondRetryMaxWait: { KeyName: "history.ReplicationTaskProcessorErrorSecondRetryMaxWait", Filters: []Filter{ShardID}, Description: "ReplicationTaskProcessorErrorSecondRetryMaxWait is the max wait time for the second phase retry", DefaultValue: time.Second * 30, }, ReplicationTaskProcessorErrorSecondRetryExpiration: { KeyName: "history.ReplicationTaskProcessorErrorSecondRetryExpiration", Filters: []Filter{ShardID}, Description: "ReplicationTaskProcessorErrorSecondRetryExpiration is the expiration duration for the second phase retry", DefaultValue: time.Minute * 5, }, ReplicationTaskProcessorNoTaskInitialWait: { KeyName: "history.ReplicationTaskProcessorNoTaskInitialWait", Filters: []Filter{ShardID}, Description: "ReplicationTaskProcessorNoTaskInitialWait is the wait time when not ask is returned", DefaultValue: time.Second * 2, }, ReplicationTaskProcessorCleanupInterval: { KeyName: "history.ReplicationTaskProcessorCleanupInterval", Filters: []Filter{ShardID}, Description: "ReplicationTaskProcessorCleanupInterval determines how frequently the cleanup replication queue", DefaultValue: time.Minute, }, ReplicationTaskProcessorStartWait: { KeyName: "history.ReplicationTaskProcessorStartWait", Filters: []Filter{ShardID}, Description: "ReplicationTaskProcessorStartWait is the wait time before each task processing batch", DefaultValue: time.Second * 5, }, WorkerESProcessorFlushInterval: { KeyName: "worker.ESProcessorFlushInterval", Description: "WorkerESProcessorFlushInterval is flush interval for esProcessor", DefaultValue: time.Second, }, WorkerTimeLimitPerArchivalIteration: { KeyName: "worker.TimeLimitPerArchivalIteration", Description: "WorkerTimeLimitPerArchivalIteration is controls the time limit of each iteration of archival workflow", DefaultValue: time.Hour * 24 * 15, }, WorkerReplicationTaskMaxRetryDuration: { KeyName: "worker.replicationTaskMaxRetryDuration", Description: "WorkerReplicationTaskMaxRetryDuration is the max retry duration for any task", DefaultValue: time.Minute * 10, }, ESAnalyzerTimeWindow: { KeyName: "worker.ESAnalyzerTimeWindow", Description: "ESAnalyzerTimeWindow defines the time window ElasticSearch Analyzer will consider while taking workflow averages", DefaultValue: time.Hour * 24 * 30, }, IsolationGroupStateRefreshInterval: { KeyName: "system.isolationGroupStateRefreshInterval", Description: "the frequency by which the IsolationGroupState handler will poll configuration", DefaultValue: time.Second * 30, }, IsolationGroupStateFetchTimeout: { KeyName: "system.IsolationGroupStateFetchTimeout", Description: "IsolationGroupStateFetchTimeout is the dynamic config DB fetch timeout value", DefaultValue: time.Second * 30, }, IsolationGroupStateUpdateTimeout: { KeyName: "system.IsolationGroupStateUpdateTimeout", Description: "IsolationGroupStateFetchTimeout is the dynamic config DB update timeout value", DefaultValue: time.Second * 30, }, ESAnalyzerBufferWaitTime: { KeyName: "worker.ESAnalyzerBufferWaitTime", Description: "ESAnalyzerBufferWaitTime controls min time required to consider a worklow stuck", DefaultValue: time.Minute * 30, }, AsyncTaskDispatchTimeout: { KeyName: "matching.asyncTaskDispatchTimeout", Filters: []Filter{DomainName, TaskListName, TaskType}, Description: "AsyncTaskDispatchTimeout is the timeout of dispatching tasks for async match", DefaultValue: time.Second * 3, }, HistoryGlobalRatelimiterDecayAfter: { KeyName: "history.globalRatelimiterDecayAfter", Description: "HistoryGlobalRatelimiterDecayAfter defines how long to wait for an update before considering a host's data \"possibly gone\", causing its weight to gradually decline.", DefaultValue: 6 * time.Second, }, HistoryGlobalRatelimiterGCAfter: { KeyName: "history.globalRatelimiterGCAfter", Description: "HistoryGlobalRatelimiterGCAfter defines how long to wait until a host's data is considered entirely useless, e.g. host has likely disappeared, its weight is very low, and the data can be deleted.", DefaultValue: 30 * time.Second, }, LocalPollWaitTime: { KeyName: "matching.localPollWaitTime", Filters: []Filter{DomainName, TaskListName, TaskType}, Description: "LocalPollWaitTime is the time a poller waits before considering request forwarding.", DefaultValue: time.Millisecond * 10, }, LocalTaskWaitTime: { KeyName: "matching.localTaskWaitTime", Filters: []Filter{DomainName, TaskListName, TaskType}, Description: "LocalTaskWaitTime is the time a task waits for a poller to arrive before considering task forwarding", DefaultValue: time.Millisecond * 10, }, }
var FloatKeys = map[FloatKey]DynamicFloat{ TestGetFloat64PropertyKey: { KeyName: "testGetFloat64PropertyKey", Description: "", DefaultValue: 0, }, TestGetFloat64PropertyFilteredByShardIDKey: { KeyName: "testGetFloat64PropertyFilteredByShardIDKey", Description: "", DefaultValue: 0, Filters: nil, }, PersistenceErrorInjectionRate: { KeyName: "system.persistenceErrorInjectionRate", Description: "PersistenceErrorInjectionRate is rate for injecting random error in persistence", DefaultValue: 0, }, AdminErrorInjectionRate: { KeyName: "admin.errorInjectionRate", Description: "dminErrorInjectionRate is the rate for injecting random error in admin client", DefaultValue: 0, }, DomainFailoverRefreshTimerJitterCoefficient: { KeyName: "frontend.domainFailoverRefreshTimerJitterCoefficient", Description: "DomainFailoverRefreshTimerJitterCoefficient is the jitter for domain failover refresh timer jitter", DefaultValue: 0.1, }, FrontendErrorInjectionRate: { KeyName: "frontend.errorInjectionRate", Description: "FrontendErrorInjectionRate is rate for injecting random error in frontend client", DefaultValue: 0, }, MatchingErrorInjectionRate: { KeyName: "matching.errorInjectionRate", Description: "MatchingErrorInjectionRate is rate for injecting random error in matching client", DefaultValue: 0, }, TaskRedispatchIntervalJitterCoefficient: { KeyName: "history.taskRedispatchIntervalJitterCoefficient", Description: "TaskRedispatchIntervalJitterCoefficient is the task redispatch interval jitter coefficient", DefaultValue: 0.15, }, QueueProcessorRandomSplitProbability: { KeyName: "history.queueProcessorRandomSplitProbability", Description: "QueueProcessorRandomSplitProbability is the probability for a domain to be split to a new processing queue", DefaultValue: 0.01, }, QueueProcessorPollBackoffIntervalJitterCoefficient: { KeyName: "history.queueProcessorPollBackoffIntervalJitterCoefficient", Description: "QueueProcessorPollBackoffIntervalJitterCoefficient is backoff interval jitter coefficient", DefaultValue: 0.15, }, TimerProcessorUpdateAckIntervalJitterCoefficient: { KeyName: "history.timerProcessorUpdateAckIntervalJitterCoefficient", Description: "TimerProcessorUpdateAckIntervalJitterCoefficient is the update interval jitter coefficient", DefaultValue: 0.15, }, TimerProcessorMaxPollIntervalJitterCoefficient: { KeyName: "history.timerProcessorMaxPollIntervalJitterCoefficient", Description: "TimerProcessorMaxPollIntervalJitterCoefficient is the max poll interval jitter coefficient", DefaultValue: 0.15, }, TimerProcessorSplitQueueIntervalJitterCoefficient: { KeyName: "history.timerProcessorSplitQueueIntervalJitterCoefficient", Description: "TimerProcessorSplitQueueIntervalJitterCoefficient is the split processing queue interval jitter coefficient", DefaultValue: 0.15, }, TransferProcessorMaxPollIntervalJitterCoefficient: { KeyName: "history.transferProcessorMaxPollIntervalJitterCoefficient", Description: "TransferProcessorMaxPollIntervalJitterCoefficient is the max poll interval jitter coefficient", DefaultValue: 0.15, }, TransferProcessorSplitQueueIntervalJitterCoefficient: { KeyName: "history.transferProcessorSplitQueueIntervalJitterCoefficient", Description: "TransferProcessorSplitQueueIntervalJitterCoefficient is the split processing queue interval jitter coefficient", DefaultValue: 0.15, }, TransferProcessorUpdateAckIntervalJitterCoefficient: { KeyName: "history.transferProcessorUpdateAckIntervalJitterCoefficient", Description: "TransferProcessorUpdateAckIntervalJitterCoefficient is the update interval jitter coefficient", DefaultValue: 0.15, }, ReplicationTaskProcessorCleanupJitterCoefficient: { KeyName: "history.ReplicationTaskProcessorCleanupJitterCoefficient", Filters: []Filter{ShardID}, Description: "ReplicationTaskProcessorCleanupJitterCoefficient is the jitter for cleanup timer", DefaultValue: 0.15, }, ReplicationTaskProcessorStartWaitJitterCoefficient: { KeyName: "history.ReplicationTaskProcessorStartWaitJitterCoefficient", Filters: []Filter{ShardID}, Description: "ReplicationTaskProcessorStartWaitJitterCoefficient is the jitter for batch start wait timer", DefaultValue: 0.9, }, ReplicationTaskProcessorHostQPS: { KeyName: "history.ReplicationTaskProcessorHostQPS", Description: "ReplicationTaskProcessorHostQPS is the qps of task processing rate limiter on host level", DefaultValue: 1500, }, ReplicationTaskProcessorShardQPS: { KeyName: "history.ReplicationTaskProcessorShardQPS", Description: "ReplicationTaskProcessorShardQPS is the qps of task processing rate limiter on shard level", DefaultValue: 5, }, ReplicationTaskGenerationQPS: { KeyName: "history.ReplicationTaskGenerationQPS", Description: "ReplicationTaskGenerationQPS is the wait time between each replication task generation qps", DefaultValue: 100, }, MutableStateChecksumInvalidateBefore: { KeyName: "history.mutableStateChecksumInvalidateBefore", Description: "MutableStateChecksumInvalidateBefore is the epoch timestamp before which all checksums are to be discarded", DefaultValue: 0, }, NotifyFailoverMarkerTimerJitterCoefficient: { KeyName: "history.NotifyFailoverMarkerTimerJitterCoefficient", Description: "NotifyFailoverMarkerTimerJitterCoefficient is the jitter for failover marker notifier timer", DefaultValue: 0.15, }, HistoryErrorInjectionRate: { KeyName: "history.errorInjectionRate", Description: "HistoryErrorInjectionRate is rate for injecting random error in history client", DefaultValue: 0, }, ReplicationTaskFetcherTimerJitterCoefficient: { KeyName: "history.ReplicationTaskFetcherTimerJitterCoefficient", Description: "ReplicationTaskFetcherTimerJitterCoefficient is the jitter for fetcher timer", DefaultValue: 0.15, }, WorkerDeterministicConstructionCheckProbability: { KeyName: "worker.DeterministicConstructionCheckProbability", Description: "WorkerDeterministicConstructionCheckProbability controls the probability of running a deterministic construction check for any given archival", DefaultValue: 0.002, }, WorkerBlobIntegrityCheckProbability: { KeyName: "worker.BlobIntegrityCheckProbability", Description: "WorkerBlobIntegrityCheckProbability controls the probability of running an integrity check for any given archival", DefaultValue: 0.002, }, HistoryGlobalRatelimiterNewDataWeight: { KeyName: "history.globalRatelimiterNewDataWeight", Description: "HistoryGlobalRatelimiterNewDataWeight defines how much weight to give each host's newest data, per update. Must be between 0 and 1, higher values match new values more closely after a single update", DefaultValue: 0.5, }, }
var IntKeys = map[IntKey]DynamicInt{}/* 184 elements not displayed */
var ListKeys = map[ListKey]DynamicList{ AllIsolationGroups: { KeyName: "system.allIsolationGroups", Description: "A list of all the isolation groups in a system", }, DefaultIsolationGroupConfigStoreManagerGlobalMapping: { KeyName: "system.defaultIsolationGroupConfigStoreManagerGlobalMapping", Description: "A configuration store for global isolation groups - used in isolation-group config only, not normal dynamic config." + "Not intended for use in normal dynamic config", }, HeaderForwardingRules: { KeyName: "admin.HeaderForwardingRules", Description: "Only loaded at startup. " + "A list of rpc.HeaderRule values that define which headers to include or exclude for all requests, applied in order. " + "Regexes and header names are used as-is, you are strongly encouraged to use `(?i)` to make your regex case-insensitive.", DefaultValue: []interface{}{ map[string]interface{}{ "Add": true, "Match": "", }, }, }, }
var MapKeys = map[MapKey]DynamicMap{ TestGetMapPropertyKey: { KeyName: "testGetMapPropertyKey", Description: "", DefaultValue: nil, }, RequiredDomainDataKeys: { KeyName: "system.requiredDomainDataKeys", Description: "RequiredDomainDataKeys is the key for the list of data keys required in domain registration", DefaultValue: nil, }, ValidSearchAttributes: { KeyName: "frontend.validSearchAttributes", Description: "ValidSearchAttributes is legal indexed keys that can be used in list APIs. When overriding, ensure to include the existing default attributes of the current release", DefaultValue: definition.GetDefaultIndexedKeys(), }, TaskSchedulerRoundRobinWeights: { KeyName: "history.taskSchedulerRoundRobinWeight", Description: "TaskSchedulerRoundRobinWeights is the priority weight for weighted round robin task scheduler", DefaultValue: common.ConvertIntMapToDynamicConfigMapProperty(map[int]int{ common.GetTaskPriority(common.HighPriorityClass, common.DefaultPrioritySubclass): 500, common.GetTaskPriority(common.DefaultPriorityClass, common.DefaultPrioritySubclass): 20, common.GetTaskPriority(common.LowPriorityClass, common.DefaultPrioritySubclass): 5, }), }, QueueProcessorPendingTaskSplitThreshold: { KeyName: "history.queueProcessorPendingTaskSplitThreshold", Description: "QueueProcessorPendingTaskSplitThreshold is the threshold for the number of pending tasks per domain", DefaultValue: common.ConvertIntMapToDynamicConfigMapProperty(map[int]int{0: 1000, 1: 10000}), }, QueueProcessorStuckTaskSplitThreshold: { KeyName: "history.queueProcessorStuckTaskSplitThreshold", Description: "QueueProcessorStuckTaskSplitThreshold is the threshold for the number of attempts of a task", DefaultValue: common.ConvertIntMapToDynamicConfigMapProperty(map[int]int{0: 100, 1: 10000}), }, }
var NotFoundError = &types.EntityNotExistsError{
Message: "unable to find key",
}
var StringKeys = map[StringKey]DynamicString{ TestGetStringPropertyKey: { KeyName: "testGetStringPropertyKey", Description: "", DefaultValue: "", }, AdvancedVisibilityWritingMode: { KeyName: "system.advancedVisibilityWritingMode", Description: "AdvancedVisibilityWritingMode is key for how to write to advanced visibility. The most useful option is dual, which can be used for seamless migration from db visibility to advanced visibility, usually using with EnableReadVisibilityFromES", DefaultValue: "on", }, AdvancedVisibilityMigrationWritingMode: { KeyName: "system.advancedVisibilityMigrationWritingMode", Description: "AdvancedVisibilityMigrationWritingMode is key for how to write to advanced visibility. The most useful option is dual, which can be used for seamless migration from advanced visibility to another", DefaultValue: "dual", }, HistoryArchivalStatus: { KeyName: "system.historyArchivalStatus", Description: "HistoryArchivalStatus is key for the status of history archival to override the value from static config.", DefaultValue: "enabled", }, VisibilityArchivalStatus: { KeyName: "system.visibilityArchivalStatus", Description: "VisibilityArchivalStatus is key for the status of visibility archival to override the value from static config.", DefaultValue: "enabled", }, DefaultEventEncoding: { KeyName: "history.defaultEventEncoding", Filters: []Filter{DomainName}, Description: "DefaultEventEncoding is the encoding type for history events", DefaultValue: string(common.EncodingTypeThriftRW), }, AdminOperationToken: { KeyName: "history.adminOperationToken", Description: "AdminOperationToken is the token to pass admin checking", DefaultValue: "CadenceTeamONLY", }, ESAnalyzerLimitToTypes: { KeyName: "worker.ESAnalyzerLimitToTypes", Description: "ESAnalyzerLimitToTypes controls if we want to limit ESAnalyzer only to some workflow types", DefaultValue: "", }, ESAnalyzerLimitToDomains: { KeyName: "worker.ESAnalyzerLimitToDomains", Description: "ESAnalyzerLimitToDomains controls if we want to limit ESAnalyzer only to some domains", DefaultValue: "", }, ESAnalyzerWorkflowDurationWarnThresholds: { KeyName: "worker.ESAnalyzerWorkflowDurationWarnThresholds", Description: "ESAnalyzerWorkflowDurationWarnThresholds defines the warning execution thresholds for workflow types", DefaultValue: "", }, ESAnalyzerWorkflowVersionMetricDomains: { KeyName: "worker.ESAnalyzerWorkflowVersionMetricDomains", Description: "ESAnalyzerWorkflowDurationWarnThresholds defines the domains we want to emit wf version metrics on", DefaultValue: "", }, ESAnalyzerWorkflowTypeMetricDomains: { KeyName: "worker.ESAnalyzerWorkflowTypeMetricDomains", Description: "ESAnalyzerWorkflowDurationWarnThresholds defines the domains we want to emit wf version metrics on", DefaultValue: "", }, FrontendGlobalRatelimiterMode: { KeyName: "frontend.globalRatelimiterMode", Description: "FrontendGlobalRatelimiterMode defines which mode a global key should be in, per key, to make gradual changes to ratelimiter algorithms", DefaultValue: "disabled", Filters: []Filter{RatelimitKey}, }, TasklistLoadBalancerStrategy: { KeyName: "system.tasklistLoadBalancerStrategy", Description: "TasklistLoadBalancerStrategy is the key for tasklist load balancer strategy", DefaultValue: "random", Filters: []Filter{DomainName, TaskListName, TaskType}, }, }
Functions ¶
func GetAllKeys ¶ added in v1.0.0
GetAllKeys returns a copy of all configuration keys with all details
func GetBoolPropertyFilteredByTaskListInfo ¶
func GetBoolPropertyFilteredByTaskListInfo(value bool) func(domain string, taskList string, taskType int) bool
GetBoolPropertyFilteredByTaskListInfo returns value as BoolPropertyFnWithTaskListInfoFilters
func GetBoolPropertyFn ¶
func GetBoolPropertyFn(value bool) func(opts ...FilterOption) bool
GetBoolPropertyFn returns value as BoolPropertyFn
func GetBoolPropertyFnFilteredByDomain ¶
GetBoolPropertyFnFilteredByDomain returns value as BoolPropertyFnWithDomainFilters
func GetBoolPropertyFnFilteredByDomainID ¶ added in v0.23.1
GetBoolPropertyFnFilteredByDomainID returns value as BoolPropertyFnWithDomainIDFilters
func GetDurationPropertyFilteredByWorkflowType ¶ added in v0.24.0
func GetDurationPropertyFilteredByWorkflowType(value time.Duration) func(domainName string, workflowType string) time.Duration
GetDurationPropertyFilteredByWorkflowType returns values as IntPropertyFnWithWorkflowTypeFilters
func GetDurationPropertyFn ¶
func GetDurationPropertyFn(value time.Duration) func(opts ...FilterOption) time.Duration
GetDurationPropertyFn returns value as DurationPropertyFn
func GetDurationPropertyFnFilteredByDomain ¶
GetDurationPropertyFnFilteredByDomain returns value as DurationPropertyFnFilteredByDomain
func GetDurationPropertyFnFilteredByShardID ¶ added in v0.24.0
GetDurationPropertyFnFilteredByShardID returns value as DurationPropertyFnWithShardIDFilter
func GetDurationPropertyFnFilteredByTaskListInfo ¶
func GetDurationPropertyFnFilteredByTaskListInfo(value time.Duration) func(domain string, taskList string, taskType int) time.Duration
GetDurationPropertyFnFilteredByTaskListInfo returns value as DurationPropertyFnWithTaskListInfoFilters
func GetFloatPropertyFn ¶
func GetFloatPropertyFn(value float64) func(opts ...FilterOption) float64
GetFloatPropertyFn returns value as FloatPropertyFn
func GetIntPropertyFilteredByDomain ¶
GetIntPropertyFilteredByDomain returns values as IntPropertyFnWithDomainFilters
func GetIntPropertyFilteredByShardID ¶ added in v0.24.0
GetIntPropertyFilteredByShardID returns values as IntPropertyFnWithShardIDFilter
func GetIntPropertyFilteredByTaskListInfo ¶
func GetIntPropertyFilteredByTaskListInfo(value int) func(domain string, taskList string, taskType int) int
GetIntPropertyFilteredByTaskListInfo returns value as IntPropertyFnWithTaskListInfoFilters
func GetIntPropertyFilteredByWorkflowType ¶ added in v0.24.0
func GetIntPropertyFilteredByWorkflowType(value int) func(domainName string, workflowType string) int
GetIntPropertyFilteredByWorkflowType returns values as IntPropertyFnWithWorkflowTypeFilters
func GetIntPropertyFn ¶
func GetIntPropertyFn(value int) func(opts ...FilterOption) int
GetIntPropertyFn returns value as IntPropertyFn
func GetMapPropertyFn ¶
func GetMapPropertyFn(value map[string]interface{}) func(opts ...FilterOption) map[string]interface{}
GetMapPropertyFn returns value as MapPropertyFn
func GetStringPropertyFn ¶
func GetStringPropertyFn(value string) func(opts ...FilterOption) string
GetStringPropertyFn returns value as StringPropertyFn
func ToGetDynamicConfigFilterRequest ¶ added in v1.2.1
func ToGetDynamicConfigFilterRequest(configName string, filters []FilterOption) *types.GetDynamicConfigRequest
ToGetDynamicConfigFilterRequest generates a GetDynamicConfigRequest object by converting filters to DynamicConfigFilter objects and setting values
func ValidateKeyValuePair ¶ added in v0.25.0
Types ¶
type BoolKey ¶ added in v0.25.0
type BoolKey int
const ( UnknownBoolKey BoolKey = iota // key for tests TestGetBoolPropertyKey TestGetBoolPropertyFilteredByDomainIDKey TestGetBoolPropertyFilteredByTaskListInfoKey TestGetBoolPropertyFilteredByDomainKey TestGetBoolPropertyFilteredByDomainIDAndWorkflowIDKey // EnableVisibilitySampling is key for enable visibility sampling for basic(DB based) visibility // KeyName: system.enableVisibilitySampling // Value type: Bool // Default value: false // Allowed filters: N/A EnableVisibilitySampling // EnableReadFromClosedExecutionV2 is key for enable read from cadence_visibility.closed_executions_v2 // KeyName: system.enableReadFromClosedExecutionV2 // Value type: Bool // Default value: false // Allowed filters: N/A EnableReadFromClosedExecutionV2 // EnableReadVisibilityFromES is key for enable read from elastic search or db visibility, usually using with AdvancedVisibilityWritingMode for seamless migration from db visibility to advanced visibility // KeyName: system.enableReadVisibilityFromES // Value type: Bool // Default value: true // Allowed filters: DomainName EnableReadVisibilityFromES // EnableReadVisibilityFromPinot is key for enable read from pinot or db visibility, usually using with AdvancedVisibilityWritingMode for seamless migration from db visibility to advanced visibility // KeyName: system.enableReadVisibilityFromPinot // Value type: Bool // Default value: true // Allowed filters: DomainName EnableReadVisibilityFromPinot // EnableVisibilityDoubleRead is the key for enable double read for a latency comparison // KeyName: system.EnableVisibilityDoubleRead // Value type: Bool // Default value: false // Allowed filters: DomainName EnableVisibilityDoubleRead // EnableLogCustomerQueryParameter is key for enable log customer query parameters // KeyName: system.enableLogCustomerQueryParameter // Value type: Bool // Default value: false EnableLogCustomerQueryParameter // EmitShardDiffLog is whether emit the shard diff log // KeyName: history.emitShardDiffLog // Value type: Bool // Default value: false // Allowed filters: N/A EmitShardDiffLog // DisableListVisibilityByFilter is config to disable list open/close workflow using filter // KeyName: frontend.disableListVisibilityByFilter // Value type: Bool // Default value: false // Allowed filters: DomainName DisableListVisibilityByFilter // EnableReadFromHistoryArchival is key for enabling reading history from archival store // KeyName: system.enableReadFromHistoryArchival // Value type: Bool // Default value: true // Allowed filters: N/A EnableReadFromHistoryArchival // EnableReadFromVisibilityArchival is key for enabling reading visibility from archival store to override the value from static config. // KeyName: system.enableReadFromVisibilityArchival // Value type: Bool // Default value: true // Allowed filters: N/A EnableReadFromVisibilityArchival // EnableDomainNotActiveAutoForwarding decides requests form which domain will be forwarded to active cluster if domain is not active in current cluster. // Only when "selected-api-forwarding" or "all-domain-apis-forwarding" is the policy in ClusterRedirectionPolicy(in static config). // If the policy is "noop"(default) this flag is not doing anything. // KeyName: system.enableDomainNotActiveAutoForwarding // Value type: Bool // Default value: true (meaning all domains are allowed to use the policy specified in static config) // Allowed filters: DomainName EnableDomainNotActiveAutoForwarding // EnableGracefulFailover is whether enabling graceful failover // KeyName: system.enableGracefulFailover // Value type: Bool // Default value: true // Allowed filters: N/A EnableGracefulFailover // DisallowQuery is the key to disallow query for a domain // KeyName: system.disallowQuery // Value type: Bool // Default value: false // Allowed filters: DomainName DisallowQuery // EnableDebugMode is for enabling debugging components, logs and metrics // KeyName: system.enableDebugMode // Value type: Bool // Default value: false // Allowed filters: N/A EnableDebugMode // EnableGRPCOutbound is the key for enabling outbound GRPC traffic // KeyName: system.enableGRPCOutbound // Value type: Bool // Default value: true // Allowed filters: N/A EnableGRPCOutbound // EnableSQLAsyncTransaction is the key for enabling async transaction // KeyName: system.enableSQLAsyncTransaction // Value type: Bool // Default value: false // Allowed filters: N/A EnableSQLAsyncTransaction // EnableConnectionRetainingDirectChooser is the key for enabling connection retaining direct yarpc chooser // KeyName: system.enableConnectionRetainingDirectChooser // Value type: Bool // Default value: false // Allowed filters: N/A EnableConnectionRetainingDirectChooser // EnableClientVersionCheck is enables client version check for frontend // KeyName: frontend.enableClientVersionCheck // Value type: Bool // Default value: false // Allowed filters: N/A EnableClientVersionCheck // SendRawWorkflowHistory is whether to enable raw history retrieving // KeyName: frontend.sendRawWorkflowHistory // Value type: Bool // Default value: false // Allowed filters: DomainName SendRawWorkflowHistory // FrontendEmitSignalNameMetricsTag enables emitting signal name tag in metrics in frontend client // KeyName: frontend.emitSignalNameMetricsTag // Value type: Bool // Default value: false // Allowed filters: DomainName FrontendEmitSignalNameMetricsTag // EnableQueryAttributeValidation enables validation of queries' search attributes against the dynamic config whitelist // Keyname: frontend.enableQueryAttributeValidation // Value type: Bool // Default value: true // Allowed filters: N/A EnableQueryAttributeValidation // MatchingEnableSyncMatch is to enable sync match // KeyName: matching.enableSyncMatch // Value type: Bool // Default value: true // Allowed filters: DomainName,TasklistName,TasklistType MatchingEnableSyncMatch // MatchingEnableTaskInfoLogByDomainID is enables info level logs for decision/activity task based on the request domainID // KeyName: matching.enableTaskInfoLogByDomainID // Value type: Bool // Default value: false // Allowed filters: DomainID MatchingEnableTaskInfoLogByDomainID // MatchingEnableTasklistGuardAgainstOwnershipShardLoss // enables guards to prevent tasklists from processing if there is any detection that the host // no longer is active or owns the shard // KeyName: matching.enableTasklistGuardAgainstOwnershipLoss // Value type: Bool // Default value: false MatchingEnableTasklistGuardAgainstOwnershipShardLoss MatchingEnableGetNumberOfPartitionsFromCache // EventsCacheGlobalEnable is enables global cache over all history shards // KeyName: history.eventsCacheGlobalEnable // Value type: Bool // Default value: false // Allowed filters: N/A EventsCacheGlobalEnable // QueueProcessorEnableSplit indicates whether processing queue split policy should be enabled // KeyName: history.queueProcessorEnableSplit // Value type: Bool // Default value: false // Allowed filters: N/A QueueProcessorEnableSplit // QueueProcessorEnableRandomSplitByDomainID indicates whether random queue split policy should be enabled for a domain // KeyName: history.queueProcessorEnableRandomSplitByDomainID // Value type: Bool // Default value: false // Allowed filters: DomainID QueueProcessorEnableRandomSplitByDomainID // QueueProcessorEnablePendingTaskSplitByDomainID indicates whether pending task split policy should be enabled // KeyName: history.queueProcessorEnablePendingTaskSplitByDomainID // Value type: Bool // Default value: false // Allowed filters: DomainID QueueProcessorEnablePendingTaskSplitByDomainID // QueueProcessorEnableStuckTaskSplitByDomainID indicates whether stuck task split policy should be enabled // KeyName: history.queueProcessorEnableStuckTaskSplitByDomainID // Value type: Bool // Default value: false // Allowed filters: DomainID QueueProcessorEnableStuckTaskSplitByDomainID // QueueProcessorEnablePersistQueueStates indicates whether processing queue states should be persisted // KeyName: history.queueProcessorEnablePersistQueueStates // Value type: Bool // Default value: true // Allowed filters: N/A QueueProcessorEnablePersistQueueStates // QueueProcessorEnableLoadQueueStates indicates whether processing queue states should be loaded // KeyName: history.queueProcessorEnableLoadQueueStates // Value type: Bool // Default value: true // Allowed filters: N/A QueueProcessorEnableLoadQueueStates // QueueProcessorEnableGracefulSyncShutdown indicates whether processing queue should be shutdown gracefully & synchronously // KeyName: history.queueProcessorEnableGracefulSyncShutdown // Value type: Bool // Default value: false // Allowed filters: N/A QueueProcessorEnableGracefulSyncShutdown // ReplicationTaskFetcherEnableGracefulSyncShutdown indicates whether task fetcher should be shutdown gracefully & synchronously // KeyName: history.replicationTaskFetcherEnableGracefulSyncShutdown // Value type: Bool // Default value: false // Allowed filters: N/A ReplicationTaskFetcherEnableGracefulSyncShutdown // TransferProcessorEnableValidator is whether validator should be enabled for transferQueueProcessor // KeyName: history.transferProcessorEnableValidator // Value type: Bool // Default value: false // Allowed filters: N/A TransferProcessorEnableValidator // EnableAdminProtection is whether to enable admin checking // KeyName: history.enableAdminProtection // Value type: Bool // Default value: false // Allowed filters: N/A EnableAdminProtection // EnableParentClosePolicy is whether to ParentClosePolicy // KeyName: history.enableParentClosePolicy // Value type: Bool // Default value: true // Allowed filters: DomainName EnableParentClosePolicy // EnableDropStuckTaskByDomainID is whether stuck timer/transfer task should be dropped for a domain // KeyName: history.DropStuckTaskByDomain // Value type: Bool // Default value: false // Allowed filters: DomainID EnableDropStuckTaskByDomainID // EnableConsistentQuery indicates if consistent query is enabled for the cluster // KeyName: history.EnableConsistentQuery // Value type: Bool // Default value: true // Allowed filters: N/A EnableConsistentQuery // EnableConsistentQueryByDomain indicates if consistent query is enabled for a domain // KeyName: history.EnableConsistentQueryByDomain // Value type: Bool // Default value: false // Allowed filters: DomainName EnableConsistentQueryByDomain // EnableContextHeaderInVisibility is key for enable context header in visibility // KeyName: history.enableContextHeaderInVisibility // Value type: Bool // Default value: false // Allowed filters: DomainName EnableContextHeaderInVisibility // EnableCrossClusterOperationsForDomain indicates if cross cluster operations can be scheduled for a domain // KeyName: history.enableCrossClusterOperations // Value type: Bool // Default value: false // Allowed filters: DomainName EnableCrossClusterOperationsForDomain // EnableHistoryCorruptionCheck enables additional sanity check for corrupted history. This allows early catches of DB corruptions but potiantally increased latency. // KeyName: history.enableHistoryCorruptionCheck // Value type: Bool // Default value: false // Allowed filters: DomainName EnableHistoryCorruptionCheck // EnableActivityLocalDispatchByDomain is allows worker to dispatch activity tasks through local tunnel after decisions are made. This is an performance optimization to skip activity scheduling efforts // KeyName: history.enableActivityLocalDispatchByDomain // Value type: Bool // Default value: true // Allowed filters: DomainName EnableActivityLocalDispatchByDomain // HistoryEnableTaskInfoLogByDomainID is enables info level logs for decision/activity task based on the request domainID // KeyName: history.enableTaskInfoLogByDomainID // Value type: Bool // Default value: false // Allowed filters: DomainID HistoryEnableTaskInfoLogByDomainID // EnableReplicationTaskGeneration is the flag to control replication generation // KeyName: history.enableReplicationTaskGeneration // Value type: Bool // Default value: true // Allowed filters: DomainID, WorkflowID EnableReplicationTaskGeneration // UseNewInitialFailoverVersion is a switch to issue a failover version based on the minFailoverVersion // rather than the default initialFailoverVersion. USed as a per-domain migration switch // KeyName: history.useNewInitialFailoverVersion // Value type: Bool // Default value: false // Allowed filters: N/A UseNewInitialFailoverVersion // EnableRecordWorkflowExecutionUninitialized enables record workflow execution uninitialized state in ElasticSearch // KeyName: history.EnableRecordWorkflowExecutionUninitialized // Value type: Bool // Default value: true // Allowed filters: DomainName EnableRecordWorkflowExecutionUninitialized // WorkflowIDCacheExternalEnabled is the key to enable/disable caching of workflowID specific information for external requests // KeyName: history.workflowIDCacheExternalEnabled // Value type: Bool // Default value: false // Allowed filters: DomainName WorkflowIDCacheExternalEnabled // WorkflowIDCacheInternalEnabled is the key to enable/disable caching of workflowID specific information for internal requests // KeyName: history.workflowIDCacheInternalEnabled // Value type: Bool // Default value: false // Allowed filters: DomainName WorkflowIDCacheInternalEnabled // WorkflowIDExternalRateLimitEnabled is the key to enable/disable rate limiting for workflowID specific information for external requests // KeyName: history.workflowIDExternalRateLimitEnabled // Value type: Bool // Default value: false // Allowed filters: DomainName WorkflowIDExternalRateLimitEnabled // WorkflowIDInternalRateLimitEnabled is the key to enable/disable rate limiting for workflowID specific information for internal requests // KeyName: history.workflowIDInternalRateLimitEnabled // Value type: Bool // Default value: false // Allowed filters: DomainName WorkflowIDInternalRateLimitEnabled // AllowArchivingIncompleteHistory will continue on when seeing some error like history mutated(usually caused by database consistency issues) // KeyName: worker.AllowArchivingIncompleteHistory // Value type: Bool // Default value: false // Allowed filters: N/A AllowArchivingIncompleteHistory // EnableCleaningOrphanTaskInTasklistScavenger indicates if enabling the scanner to clean up orphan tasks // Only implemented for single SQL database. TODO https://github.com/uber/cadence/issues/4064 for supporting multiple/sharded SQL database and NoSQL // KeyName: worker.enableCleaningOrphanTaskInTasklistScavenger // Value type: Bool // Default value: false // Allowed filters: N/A EnableCleaningOrphanTaskInTasklistScavenger // TaskListScannerEnabled indicates if task list scanner should be started as part of worker.Scanner // KeyName: worker.taskListScannerEnabled // Value type: Bool // Default value: true // Allowed filters: N/A TaskListScannerEnabled // HistoryScannerEnabled indicates if history scanner should be started as part of worker.Scanner // KeyName: worker.historyScannerEnabled // Value type: Bool // Default value: false // Allowed filters: N/A HistoryScannerEnabled // ConcreteExecutionsScannerEnabled indicates if executions scanner should be started as part of worker.Scanner // KeyName: worker.executionsScannerEnabled // Value type: Bool // Default value: false // Allowed filters: N/A ConcreteExecutionsScannerEnabled // ConcreteExecutionsScannerInvariantCollectionMutableState indicates if mutable state invariant checks should be run // KeyName: worker.executionsScannerInvariantCollectionMutableState // Value type: Bool // Default value: true // Allowed filters: N/A ConcreteExecutionsScannerInvariantCollectionMutableState // ConcreteExecutionsScannerInvariantCollectionHistory indicates if history invariant checks should be run // KeyName: worker.executionsScannerInvariantCollectionHistory // Value type: Bool // Default value: true // Allowed filters: N/A ConcreteExecutionsScannerInvariantCollectionHistory // ConcreteExecutionsFixerInvariantCollectionStale indicates if the stale workflow invariant should be run // KeyName: worker.executionsFixerInvariantCollectionStale // Value type: Bool // Default value: false // Allowed filters: N/A ConcreteExecutionsFixerInvariantCollectionStale // ConcreteExecutionsFixerInvariantCollectionMutableState indicates if mutable state invariant checks should be run // KeyName: worker.executionsFixerInvariantCollectionMutableState // Value type: Bool // Default value: true // Allowed filters: N/A ConcreteExecutionsFixerInvariantCollectionMutableState // ConcreteExecutionsFixerInvariantCollectionHistory indicates if history invariant checks should be run // KeyName: worker.executionsFixerInvariantCollectionHistory // Value type: Bool // Default value: true // Allowed filters: N/A ConcreteExecutionsFixerInvariantCollectionHistory // ConcreteExecutionsScannerInvariantCollectionStale indicates if the stale workflow invariant should be run // KeyName: worker.executionsScannerInvariantCollectionStale // Value type: Bool // Default value: false // Allowed filters: N/A ConcreteExecutionsScannerInvariantCollectionStale // CurrentExecutionsScannerEnabled indicates if current executions scanner should be started as part of worker.Scanner // KeyName: worker.currentExecutionsScannerEnabled // Value type: Bool // Default value: false // Allowed filters: N/A CurrentExecutionsScannerEnabled // CurrentExecutionsScannerInvariantCollectionHistory indicates if history invariant checks should be run // KeyName: worker.currentExecutionsScannerInvariantCollectionHistory // Value type: Bool // Default value: true // Allowed filters: N/A CurrentExecutionsScannerInvariantCollectionHistory // CurrentExecutionsScannerInvariantCollectionMutableState indicates if mutable state invariant checks should be run // KeyName: worker.currentExecutionsInvariantCollectionMutableState // Value type: Bool // Default value: true // Allowed filters: N/A CurrentExecutionsScannerInvariantCollectionMutableState // EnableBatcher is decides whether start batcher in our worker // KeyName: worker.enableBatcher // Value type: Bool // Default value: true // Allowed filters: N/A EnableBatcher // EnableParentClosePolicyWorker decides whether or not enable system workers for processing parent close policy task // KeyName: system.enableParentClosePolicyWorker // Value type: Bool // Default value: true // Allowed filters: N/A EnableParentClosePolicyWorker // EnableESAnalyzer decides whether to enable system workers for processing ElasticSearch Analyzer // KeyName: system.enableESAnalyzer // Value type: Bool // Default value: false // Allowed filters: N/A EnableESAnalyzer // EnableAsyncWorkflowConsumption decides whether to enable system workers for processing async workflows // KeyName: worker.enableAsyncWorkflowConsumption // Value type: Bool // Default value: false // Allowed filters: N/A EnableAsyncWorkflowConsumption // EnableStickyQuery indicates if sticky query should be enabled per domain // KeyName: system.enableStickyQuery // Value type: Bool // Default value: true // Allowed filters: DomainName EnableStickyQuery // EnableFailoverManager indicates if failover manager is enabled // KeyName: system.enableFailoverManager // Value type: Bool // Default value: true // Allowed filters: N/A EnableFailoverManager // ConcreteExecutionFixerDomainAllow is which domains are allowed to be fixed by concrete fixer workflow // KeyName: worker.concreteExecutionFixerDomainAllow // Value type: Bool // Default value: false // Allowed filters: DomainName ConcreteExecutionFixerDomainAllow // CurrentExecutionFixerDomainAllow is which domains are allowed to be fixed by current fixer workflow // KeyName: worker.currentExecutionFixerDomainAllow // Value type: Bool // Default value: false // Allowed filters: DomainName CurrentExecutionFixerDomainAllow // TimersScannerEnabled is if timers scanner should be started as part of worker.Scanner // KeyName: worker.timersScannerEnabled // Value type: Bool // Default value: false // Allowed filters: N/A TimersScannerEnabled // TimersFixerEnabled is if timers fixer should be started as part of worker.Scanner // KeyName: worker.timersFixerEnabled // Value type: Bool // Default value: false // Allowed filters: N/A TimersFixerEnabled // TimersFixerDomainAllow is which domains are allowed to be fixed by timer fixer workflow // KeyName: worker.timersFixerDomainAllow // Value type: Bool // Default value: false // Allowed filters: DomainName TimersFixerDomainAllow // ConcreteExecutionFixerEnabled is if concrete execution fixer workflow is enabled // KeyName: worker.concreteExecutionFixerEnabled // Value type: Bool // Default value: false // Allowed filters: N/A ConcreteExecutionFixerEnabled // CurrentExecutionFixerEnabled is if current execution fixer workflow is enabled // KeyName: worker.currentExecutionFixerEnabled // Value type: Bool // Default value: false // Allowed filters: N/A CurrentExecutionFixerEnabled // EnableAuthorization is the key to enable authorization for a domain, only for extension binary: // KeyName: N/A // Default value: N/A // TODO: https://github.com/uber/cadence/issues/3861 EnableAuthorization // EnableServiceAuthorization is the key to enable authorization for a service, only for extension binary: // KeyName: N/A // Default value: N/A // TODO: https://github.com/uber/cadence/issues/3861 EnableServiceAuthorization // EnableServiceAuthorizationLogOnly is the key to enable authorization logging for a service, only for extension binary: // KeyName: N/A // Default value: N/A // TODO: https://github.com/uber/cadence/issues/3861 EnableServiceAuthorizationLogOnly // ESAnalyzerPause defines if we want to dynamically pause the analyzer workflow // KeyName: worker.ESAnalyzerPause // Value type: bool // Default value: false ESAnalyzerPause // EnableArchivalCompression indicates whether blobs are compressed before they are archived // KeyName: N/A // Default value: N/A // TODO: https://github.com/uber/cadence/issues/3861 EnableArchivalCompression // ESAnalyzerEnableAvgDurationBasedChecks controls if we want to enable avg duration based task refreshes // KeyName: worker.ESAnalyzerEnableAvgDurationBasedChecks // Value type: Bool // Default value: false ESAnalyzerEnableAvgDurationBasedChecks // Lockdown defines if we want to allow failovers of domains to this cluster // KeyName: system.Lockdown // Value type: bool // Default value: false Lockdown // PendingActivityValidationEnabled is feature flag if pending activity count validation is enabled // KeyName: limit.pendingActivityCount.enabled // Value type: bool // Default value: false EnablePendingActivityValidation EnableCassandraAllConsistencyLevelDelete // EnableTasklistIsolation Is a feature to enable subdivision of workflows by units called 'isolation-groups' // and to control their movement and blast radius. This has some nontrivial operational overheads in management // and a good understanding of poller distribution, so probably not worth enabling unless it's well understood. // KeyName: system.enableTasklistIsolation // Value type: bool // Default value: false EnableTasklistIsolation // EnableShardIDMetrics turns on or off shardId metrics // KeyName: system.enableShardIDMetrics // Value type: Bool // Default value: true EnableShardIDMetrics EnableTimerDebugLogByDomainID // EnableTaskVal is which allows the taskvalidation to be enabled. // KeyName: system.enableTaskVal // Value type: Bool // Default value: false // Allowed filters: DomainID EnableTaskVal // EnableRetryForChecksumFailure enables retry if mutable state checksum verification fails // KeyName: history.enableMutableStateChecksumFailureRetry // Value type: Bool // Default value: false // Allowed filters: DomainName EnableRetryForChecksumFailure // EnableStrongIdempotency enables strong idempotency for APIs // KeyName: history.enableStrongIdempotency // Value type: Bool // Default value: false // Allowed filters: DomainName EnableStrongIdempotency // EnableStrongIdempotencySanityCheck enables sanity check for strong idempotency // KeyName: history.enableStrongIdempotencySanityCheck // Value type: Bool // Default value: false // Allowed filters: DomainName EnableStrongIdempotencySanityCheck // LastBoolKey must be the last one in this const group LastBoolKey )
func (BoolKey) DefaultBool ¶ added in v0.25.0
func (BoolKey) DefaultValue ¶ added in v0.25.0
func (k BoolKey) DefaultValue() interface{}
func (BoolKey) Description ¶ added in v0.25.0
type BoolPropertyFn ¶
type BoolPropertyFn func(opts ...FilterOption) bool
BoolPropertyFn is a wrapper to get bool property from dynamic config
type BoolPropertyFnWithDomainFilter ¶
BoolPropertyFnWithDomainFilter is a wrapper to get bool property from dynamic config with domain as filter
type BoolPropertyFnWithDomainIDAndWorkflowIDFilter ¶ added in v0.24.0
BoolPropertyFnWithDomainIDAndWorkflowIDFilter is a wrapper to get bool property from dynamic config with domainID and workflowID as filter
type BoolPropertyFnWithDomainIDFilter ¶
BoolPropertyFnWithDomainIDFilter is a wrapper to get bool property from dynamic config with domainID as filter
type BoolPropertyFnWithTaskListInfoFilters ¶
BoolPropertyFnWithTaskListInfoFilters is a wrapper to get bool property from dynamic config with three filters: domain, taskList, taskType
type Client ¶
type Client interface { GetValue(name Key) (interface{}, error) GetValueWithFilters(name Key, filters map[Filter]interface{}) (interface{}, error) GetIntValue(name IntKey, filters map[Filter]interface{}) (int, error) GetFloatValue(name FloatKey, filters map[Filter]interface{}) (float64, error) GetBoolValue(name BoolKey, filters map[Filter]interface{}) (bool, error) GetStringValue(name StringKey, filters map[Filter]interface{}) (string, error) GetMapValue(name MapKey, filters map[Filter]interface{}) (map[string]interface{}, error) GetDurationValue(name DurationKey, filters map[Filter]interface{}) (time.Duration, error) GetListValue(name ListKey, filters map[Filter]interface{}) ([]interface{}, error) // UpdateValue takes value as map and updates by overriding. It doesn't support update with filters. UpdateValue(name Key, value interface{}) error RestoreValue(name Key, filters map[Filter]interface{}) error ListValue(name Key) ([]*types.DynamicConfigEntry, error) }
Client allows fetching values from a dynamic configuration system NOTE: This does not have async options right now. In the interest of keeping it minimal, we can add when requirement arises.
func NewFileBasedClient ¶
func NewFileBasedClient(config *FileBasedClientConfig, logger log.Logger, doneCh chan struct{}) (Client, error)
NewFileBasedClient creates a file based client.
func NewInMemoryClient ¶
func NewInMemoryClient() Client
NewInMemoryClient creates a new in memory dynamic config client for testing purpose
type Collection ¶
type Collection struct {
// contains filtered or unexported fields
}
Collection wraps dynamic config client with a closure so that across the code, the config values can be directly accessed by calling the function without propagating the client everywhere in code
func NewCollection ¶
func NewCollection( client Client, logger log.Logger, filterOptions ...FilterOption, ) *Collection
NewCollection creates a new collection
func NewNopCollection ¶
func NewNopCollection() *Collection
NewNopCollection creates a new nop collection
func (*Collection) GetBoolProperty ¶
func (c *Collection) GetBoolProperty(key BoolKey) BoolPropertyFn
GetBoolProperty gets property and asserts that it's an bool
func (*Collection) GetBoolPropertyFilteredByDomain ¶
func (c *Collection) GetBoolPropertyFilteredByDomain(key BoolKey) BoolPropertyFnWithDomainFilter
GetBoolPropertyFilteredByDomain gets property with domain filter and asserts that it's a bool
func (*Collection) GetBoolPropertyFilteredByDomainID ¶
func (c *Collection) GetBoolPropertyFilteredByDomainID(key BoolKey) BoolPropertyFnWithDomainIDFilter
GetBoolPropertyFilteredByDomainID gets property with domainID filter and asserts that it's a bool
func (*Collection) GetBoolPropertyFilteredByDomainIDAndWorkflowID ¶ added in v0.24.0
func (c *Collection) GetBoolPropertyFilteredByDomainIDAndWorkflowID(key BoolKey) BoolPropertyFnWithDomainIDAndWorkflowIDFilter
GetBoolPropertyFilteredByDomainIDAndWorkflowID gets property with domainID and workflowID filters and asserts that it's a bool
func (*Collection) GetBoolPropertyFilteredByTaskListInfo ¶
func (c *Collection) GetBoolPropertyFilteredByTaskListInfo(key BoolKey) BoolPropertyFnWithTaskListInfoFilters
GetBoolPropertyFilteredByTaskListInfo gets property with taskListInfo as filters and asserts that it's an bool
func (*Collection) GetDurationProperty ¶
func (c *Collection) GetDurationProperty(key DurationKey) DurationPropertyFn
GetDurationProperty gets property and asserts that it's a duration
func (*Collection) GetDurationPropertyFilteredByDomain ¶
func (c *Collection) GetDurationPropertyFilteredByDomain(key DurationKey) DurationPropertyFnWithDomainFilter
GetDurationPropertyFilteredByDomain gets property with domain filter and asserts that it's a duration
func (*Collection) GetDurationPropertyFilteredByDomainID ¶
func (c *Collection) GetDurationPropertyFilteredByDomainID(key DurationKey) DurationPropertyFnWithDomainIDFilter
GetDurationPropertyFilteredByDomainID gets property with domainID filter and asserts that it's a duration
func (*Collection) GetDurationPropertyFilteredByShardID ¶
func (c *Collection) GetDurationPropertyFilteredByShardID(key DurationKey) DurationPropertyFnWithShardIDFilter
GetDurationPropertyFilteredByShardID gets property with shardID id as filter and asserts that it's a duration
func (*Collection) GetDurationPropertyFilteredByTaskListInfo ¶
func (c *Collection) GetDurationPropertyFilteredByTaskListInfo(key DurationKey) DurationPropertyFnWithTaskListInfoFilters
GetDurationPropertyFilteredByTaskListInfo gets property with taskListInfo as filters and asserts that it's a duration
func (*Collection) GetDurationPropertyFilteredByWorkflowType ¶ added in v0.24.0
func (c *Collection) GetDurationPropertyFilteredByWorkflowType(key DurationKey) DurationPropertyFnWithWorkflowTypeFilter
GetDurationPropertyFilteredByWorkflowType gets property with workflow type filter and asserts that it's a duration
func (*Collection) GetFloat64Property ¶
func (c *Collection) GetFloat64Property(key FloatKey) FloatPropertyFn
GetFloat64Property gets property and asserts that it's a float64
func (*Collection) GetFloat64PropertyFilteredByShardID ¶
func (c *Collection) GetFloat64PropertyFilteredByShardID(key FloatKey) FloatPropertyFnWithShardIDFilter
GetFloat64PropertyFilteredByShardID gets property with shardID filter and asserts that it's a float64
func (*Collection) GetIntProperty ¶
func (c *Collection) GetIntProperty(key IntKey) IntPropertyFn
GetIntProperty gets property and asserts that it's an integer
func (*Collection) GetIntPropertyFilteredByDomain ¶
func (c *Collection) GetIntPropertyFilteredByDomain(key IntKey) IntPropertyFnWithDomainFilter
GetIntPropertyFilteredByDomain gets property with domain filter and asserts that it's an integer
func (*Collection) GetIntPropertyFilteredByShardID ¶
func (c *Collection) GetIntPropertyFilteredByShardID(key IntKey) IntPropertyFnWithShardIDFilter
GetIntPropertyFilteredByShardID gets property with shardID as filter and asserts that it's an integer
func (*Collection) GetIntPropertyFilteredByTaskListInfo ¶
func (c *Collection) GetIntPropertyFilteredByTaskListInfo(key IntKey) IntPropertyFnWithTaskListInfoFilters
GetIntPropertyFilteredByTaskListInfo gets property with taskListInfo as filters and asserts that it's an integer
func (*Collection) GetIntPropertyFilteredByWorkflowType ¶ added in v0.24.0
func (c *Collection) GetIntPropertyFilteredByWorkflowType(key IntKey) IntPropertyFnWithWorkflowTypeFilter
GetIntPropertyFilteredByWorkflowType gets property with workflow type filter and asserts that it's an integer
func (*Collection) GetListProperty ¶ added in v1.0.0
func (c *Collection) GetListProperty(key ListKey) ListPropertyFn
func (*Collection) GetMapProperty ¶
func (c *Collection) GetMapProperty(key MapKey) MapPropertyFn
GetMapProperty gets property and asserts that it's a map
func (*Collection) GetProperty ¶
func (c *Collection) GetProperty(key Key) PropertyFn
GetProperty gets a interface property and returns defaultValue if property is not found
func (*Collection) GetStringProperty ¶
func (c *Collection) GetStringProperty(key StringKey) StringPropertyFn
GetStringProperty gets property and asserts that it's an string
func (*Collection) GetStringPropertyFilteredByDomain ¶
func (c *Collection) GetStringPropertyFilteredByDomain(key StringKey) StringPropertyFnWithDomainFilter
GetStringPropertyFilteredByDomain gets property with domain filter and asserts that it's a string
func (*Collection) GetStringPropertyFilteredByRatelimitKey ¶ added in v1.2.12
func (c *Collection) GetStringPropertyFilteredByRatelimitKey(key StringKey) StringPropertyWithRatelimitKeyFilter
func (*Collection) GetStringPropertyFilteredByTaskListInfo ¶ added in v1.2.14
func (c *Collection) GetStringPropertyFilteredByTaskListInfo(key StringKey) StringPropertyFnWithTaskListInfoFilters
type DurationKey ¶ added in v0.25.0
type DurationKey int
const ( UnknownDurationKey DurationKey = iota // key for tests TestGetDurationPropertyKey TestGetDurationPropertyFilteredByDomainKey TestGetDurationPropertyFilteredByTaskListInfoKey TestGetDurationPropertyFilteredByWorkflowTypeKey TestGetDurationPropertyFilteredByDomainIDKey TestGetDurationPropertyFilteredByShardID // FrontendShutdownDrainDuration is the duration of traffic drain during shutdown // KeyName: frontend.shutdownDrainDuration // Value type: Duration // Default value: 0 // Allowed filters: N/A FrontendShutdownDrainDuration // FrontendFailoverCoolDown is duration between two domain failvoers // KeyName: frontend.failoverCoolDown // Value type: Duration // Default value: 1m (one minute, see domain.FailoverCoolDown) // Allowed filters: DomainName FrontendFailoverCoolDown // DomainFailoverRefreshInterval is the domain failover refresh timer // KeyName: frontend.domainFailoverRefreshInterval // Value type: Duration // Default value: 10s (10*time.Second) // Allowed filters: N/A DomainFailoverRefreshInterval // GlobalRatelimiterUpdateInterval controls how frequently ratelimiter usage information is submitted to aggregators. // This value is shared between limiting and aggregating hosts (frontend and history). // KeyName: frontend.globalRatelimiterUpdateInterval // Value type: Duration // Default value: 3 seconds GlobalRatelimiterUpdateInterval // MatchingLongPollExpirationInterval is the long poll expiration interval in the matching service // KeyName: matching.longPollExpirationInterval // Value type: Duration // Default value: time.Minute // Allowed filters: DomainName,TasklistName,TasklistType MatchingLongPollExpirationInterval // MatchingUpdateAckInterval is the interval for update ack // KeyName: matching.updateAckInterval // Value type: Duration // Default value: 1m (1*time.Minute) // Allowed filters: DomainName,TasklistName,TasklistType MatchingUpdateAckInterval // MatchingIdleTasklistCheckInterval is the IdleTasklistCheckInterval // KeyName: matching.idleTasklistCheckInterval // Value type: Duration // Default value: 5m (5*time.Minute) // Allowed filters: DomainName,TasklistName,TasklistType MatchingIdleTasklistCheckInterval // MaxTasklistIdleTime is the max time tasklist being idle // KeyName: matching.maxTasklistIdleTime // Value type: Duration // Default value: 5m (5*time.Minute) // Allowed filters: DomainName,TasklistName,TasklistType MaxTasklistIdleTime // MatchingShutdownDrainDuration is the duration of traffic drain during shutdown // KeyName: matching.shutdownDrainDuration // Value type: Duration // Default value: 0 // Allowed filters: N/A MatchingShutdownDrainDuration // MatchingActivityTaskSyncMatchWaitTime is the amount of time activity task will wait to be sync matched // KeyName: matching.activityTaskSyncMatchWaitTime // Value type: Duration // Default value: 100ms // Allowed filters: DomainName MatchingActivityTaskSyncMatchWaitTime // HistoryLongPollExpirationInterval is the long poll expiration interval in the history service // KeyName: history.longPollExpirationInterval // Value type: Duration // Default value: 20s( time.Second*20) // Allowed filters: DomainName HistoryLongPollExpirationInterval // HistoryCacheTTL is TTL of history cache // KeyName: history.cacheTTL // Value type: Duration // Default value: 1h (time.Hour) // Allowed filters: N/A HistoryCacheTTL // HistoryShutdownDrainDuration is the duration of traffic drain during shutdown // KeyName: history.shutdownDrainDuration // Value type: Duration // Default value: 0 // Allowed filters: N/A HistoryShutdownDrainDuration // EventsCacheTTL is TTL of events cache // KeyName: history.eventsCacheTTL // Value type: Duration // Default value: 1h (time.Hour) // Allowed filters: N/A EventsCacheTTL // AcquireShardInterval is interval that timer used to acquire shard // KeyName: history.acquireShardInterval // Value type: Duration // Default value: 1m (time.Minute) // Allowed filters: N/A AcquireShardInterval // StandbyClusterDelay is the artificial delay added to standby cluster's view of active cluster's time // KeyName: history.standbyClusterDelay // Value type: Duration // Default value: 5m (5*time.Minute) // Allowed filters: N/A StandbyClusterDelay // StandbyTaskMissingEventsResendDelay is the amount of time standby cluster's will wait (if events are missing)before calling remote for missing events // KeyName: history.standbyTaskMissingEventsResendDelay // Value type: Duration // Default value: 15m (15*time.Minute) // Allowed filters: N/A StandbyTaskMissingEventsResendDelay // StandbyTaskMissingEventsDiscardDelay is the amount of time standby cluster's will wait (if events are missing)before discarding the task // KeyName: history.standbyTaskMissingEventsDiscardDelay // Value type: Duration // Default value: 25m (25*time.Minute) // Allowed filters: N/A StandbyTaskMissingEventsDiscardDelay // ActiveTaskRedispatchInterval is the active task redispatch interval // KeyName: history.activeTaskRedispatchInterval // Value type: Duration // Default value: 5s (5*time.Second) // Allowed filters: N/A ActiveTaskRedispatchInterval // StandbyTaskRedispatchInterval is the standby task redispatch interval // KeyName: history.standbyTaskRedispatchInterval // Value type: Duration // Default value: 30s (30*time.Second) // Allowed filters: N/A StandbyTaskRedispatchInterval // StandbyTaskReReplicationContextTimeout is the context timeout for standby task re-replication // KeyName: history.standbyTaskReReplicationContextTimeout // Value type: Duration // Default value: 3m (3*time.Minute) // Allowed filters: DomainID StandbyTaskReReplicationContextTimeout // ResurrectionCheckMinDelay is the minimal timer processing delay before scanning history to see // if there's a resurrected timer/activity // KeyName: history.resurrectionCheckMinDelay // Value type: Duration // Default value: 24h (24*time.Hour) // Allowed filters: DomainName ResurrectionCheckMinDelay // QueueProcessorSplitLookAheadDurationByDomainID is the look ahead duration when spliting a domain to a new processing queue // KeyName: history.queueProcessorSplitLookAheadDurationByDomainID // Value type: Duration // Default value: 20m (20*time.Minute) // Allowed filters: DomainID QueueProcessorSplitLookAheadDurationByDomainID // QueueProcessorPollBackoffInterval is the backoff duration when queue processor is throttled // KeyName: history.queueProcessorPollBackoffInterval // Value type: Duration // Default value: 5s (5*time.Second) // Allowed filters: N/A QueueProcessorPollBackoffInterval // TimerProcessorUpdateAckInterval is update interval for timer processor // KeyName: history.timerProcessorUpdateAckInterval // Value type: Duration // Default value: 30s (30*time.Second) // Allowed filters: N/A TimerProcessorUpdateAckInterval // TimerProcessorCompleteTimerInterval is complete timer interval for timer processor // KeyName: history.timerProcessorCompleteTimerInterval // Value type: Duration // Default value: 60s (60*time.Second) // Allowed filters: N/A TimerProcessorCompleteTimerInterval // TimerProcessorFailoverMaxStartJitterInterval is the max jitter interval for starting timer // failover queue processing. The actual jitter interval used will be a random duration between // 0 and the max interval so that timer failover queue across different shards won't start at // the same time // KeyName: history.timerProcessorFailoverMaxStartJitterInterval // Value type: Duration // Default value: 0s (0*time.Second) // Allowed filters: N/A TimerProcessorFailoverMaxStartJitterInterval // TimerProcessorMaxPollInterval is max poll interval for timer processor // KeyName: history.timerProcessorMaxPollInterval // Value type: Duration // Default value: 5m (5*time.Minute) // Allowed filters: N/A TimerProcessorMaxPollInterval // TimerProcessorSplitQueueInterval is the split processing queue interval for timer processor // KeyName: history.timerProcessorSplitQueueInterval // Value type: Duration // Default value: 1m (1*time.Minute) // Allowed filters: N/A TimerProcessorSplitQueueInterval // TimerProcessorArchivalTimeLimit is the upper time limit for inline history archival // KeyName: history.timerProcessorArchivalTimeLimit // Value type: Duration // Default value: 2s (2*time.Second) // Allowed filters: N/A TimerProcessorArchivalTimeLimit // TimerProcessorMaxTimeShift is the max shift timer processor can have // KeyName: history.timerProcessorMaxTimeShift // Value type: Duration // Default value: 1s (1*time.Second) // Allowed filters: N/A TimerProcessorMaxTimeShift // TransferProcessorFailoverMaxStartJitterInterval is the max jitter interval for starting transfer // failover queue processing. The actual jitter interval used will be a random duration between // 0 and the max interval so that timer failover queue across different shards won't start at // the same time // KeyName: history.transferProcessorFailoverMaxStartJitterInterval // Value type: Duration // Default value: 0s (0*time.Second) // Allowed filters: N/A TransferProcessorFailoverMaxStartJitterInterval // TransferProcessorMaxPollInterval is max poll interval for transferQueueProcessor // KeyName: history.transferProcessorMaxPollInterval // Value type: Duration // Default value: 1m (1*time.Minute) // Allowed filters: N/A TransferProcessorMaxPollInterval // TransferProcessorSplitQueueInterval is the split processing queue interval for transferQueueProcessor // KeyName: history.transferProcessorSplitQueueInterval // Value type: Duration // Default value: 1m (1*time.Minute) // Allowed filters: N/A TransferProcessorSplitQueueInterval // TransferProcessorUpdateAckInterval is update interval for transferQueueProcessor // KeyName: history.transferProcessorUpdateAckInterval // Value type: Duration // Default value: 30s (30*time.Second) // Allowed filters: N/A TransferProcessorUpdateAckInterval // TransferProcessorCompleteTransferInterval is complete timer interval for transferQueueProcessor // KeyName: history.transferProcessorCompleteTransferInterval // Value type: Duration // Default value: 60s (60*time.Second) // Allowed filters: N/A TransferProcessorCompleteTransferInterval // TransferProcessorValidationInterval is interval for performing transfer queue validation // KeyName: history.transferProcessorValidationInterval // Value type: Duration // Default value: 30s (30*time.Second) // Allowed filters: N/A TransferProcessorValidationInterval // TransferProcessorVisibilityArchivalTimeLimit is the upper time limit for archiving visibility records // KeyName: history.transferProcessorVisibilityArchivalTimeLimit // Value type: Duration // Default value: 400ms (400*time.Millisecond) // Allowed filters: N/A TransferProcessorVisibilityArchivalTimeLimit // ReplicatorUpperLatency indicates the max allowed replication latency between clusters // KeyName: history.replicatorUpperLatency // Value type: Duration // Default value: 40s (40 * time.Second) // Allowed filters: N/A ReplicatorUpperLatency // ShardUpdateMinInterval is the minimal time interval which the shard info can be updated // KeyName: history.shardUpdateMinInterval // Value type: Duration // Default value: 5m (5*time.Minute) // Allowed filters: N/A ShardUpdateMinInterval // ShardSyncMinInterval is the minimal time interval which the shard info should be sync to remote // KeyName: history.shardSyncMinInterval // Value type: Duration // Default value: 5m (5*time.Minute) // Allowed filters: N/A ShardSyncMinInterval // StickyTTL is to expire a sticky tasklist if no update more than this duration // KeyName: history.stickyTTL // Value type: Duration // Default value: time.Hour*24*365 // Allowed filters: DomainName StickyTTL // DecisionHeartbeatTimeout is for decision heartbeat // KeyName: history.decisionHeartbeatTimeout // Value type: Duration // Default value: 30m (time.Minute*30) // Allowed filters: DomainName DecisionHeartbeatTimeout // NormalDecisionScheduleToStartTimeout is scheduleToStart timeout duration for normal (non-sticky) decision task // KeyName: history.normalDecisionScheduleToStartTimeout // Value type: Duration // Default value: time.Minute*5 // Allowed filters: DomainName NormalDecisionScheduleToStartTimeout // NotifyFailoverMarkerInterval is determines the frequency to notify failover marker // KeyName: history.NotifyFailoverMarkerInterval // Value type: Duration // Default value: 5s (5*time.Second) // Allowed filters: N/A NotifyFailoverMarkerInterval // ActivityMaxScheduleToStartTimeoutForRetry is maximum value allowed when overwritting the schedule to start timeout for activities with retry policy // KeyName: history.activityMaxScheduleToStartTimeoutForRetry // Value type: Duration // Default value: 30m (30*time.Minute) // Allowed filters: DomainName ActivityMaxScheduleToStartTimeoutForRetry // ReplicationTaskFetcherAggregationInterval determines how frequently the fetch requests are sent // KeyName: history.ReplicationTaskFetcherAggregationInterval // Value type: Duration // Default value: 2s (2 * time.Second) // Allowed filters: N/A ReplicationTaskFetcherAggregationInterval // ReplicationTaskFetcherErrorRetryWait is the wait time when fetcher encounters error // KeyName: history.ReplicationTaskFetcherErrorRetryWait // Value type: Duration // Default value: time.Second // Allowed filters: N/A ReplicationTaskFetcherErrorRetryWait // ReplicationTaskFetcherServiceBusyWait is the wait time when fetcher encounters service busy error // KeyName: history.ReplicationTaskFetcherServiceBusyWait // Value type: Duration // Default value: 60s (60 * time.Second) // Allowed filters: N/A ReplicationTaskFetcherServiceBusyWait // ReplicationTaskProcessorErrorRetryWait is the initial retry wait when we see errors in applying replication tasks // KeyName: history.ReplicationTaskProcessorErrorRetryWait // Value type: Duration // Default value: 50ms (50*time.Millisecond) // Allowed filters: ShardID ReplicationTaskProcessorErrorRetryWait // ReplicationTaskProcessorErrorSecondRetryWait is the initial retry wait for the second phase retry // KeyName: history.ReplicationTaskProcessorErrorSecondRetryWait // Value type: Duration // Default value: 5s (5* time.Second) // Allowed filters: ShardID ReplicationTaskProcessorErrorSecondRetryWait // ReplicationTaskProcessorErrorSecondRetryMaxWait is the max wait time for the second phase retry // KeyName: history.ReplicationTaskProcessorErrorSecondRetryMaxWait // Value type: Duration // Default value: 30s (30 * time.Second) // Allowed filters: ShardID ReplicationTaskProcessorErrorSecondRetryMaxWait // ReplicationTaskProcessorErrorSecondRetryExpiration is the expiration duration for the second phase retry // KeyName: history.ReplicationTaskProcessorErrorSecondRetryExpiration // Value type: Duration // Default value: 5m (5* time.Minute) // Allowed filters: ShardID ReplicationTaskProcessorErrorSecondRetryExpiration // ReplicationTaskProcessorNoTaskInitialWait is the wait time when not ask is returned // KeyName: history.ReplicationTaskProcessorNoTaskInitialWait // Value type: Duration // Default value: 2s (2* time.Second) // Allowed filters: ShardID ReplicationTaskProcessorNoTaskInitialWait // ReplicationTaskProcessorCleanupInterval determines how frequently the cleanup replication queue // KeyName: history.ReplicationTaskProcessorCleanupInterval // Value type: Duration // Default value: 1m (1* time.Minute) // Allowed filters: ShardID ReplicationTaskProcessorCleanupInterval // ReplicationTaskProcessorStartWait is the wait time before each task processing batch // KeyName: history.ReplicationTaskProcessorStartWait // Value type: Duration // Default value: 5s (5* time.Second) // Allowed filters: ShardID ReplicationTaskProcessorStartWait // WorkerESProcessorFlushInterval is flush interval for esProcessor // KeyName: worker.ESProcessorFlushInterval // Value type: Duration // Default value: 1s (1*time.Second) // Allowed filters: N/A WorkerESProcessorFlushInterval // WorkerTimeLimitPerArchivalIteration is controls the time limit of each iteration of archival workflow // KeyName: worker.TimeLimitPerArchivalIteration // Value type: Duration // Default value: archiver.MaxArchivalIterationTimeout() // Allowed filters: N/A WorkerTimeLimitPerArchivalIteration // WorkerReplicationTaskMaxRetryDuration is the max retry duration for any task // KeyName: worker.replicationTaskMaxRetryDuration // Value type: Duration // Default value: 10m (time.Minute*10) // Allowed filters: N/A WorkerReplicationTaskMaxRetryDuration // ESAnalyzerTimeWindow defines the time window ElasticSearch Analyzer will consider while taking workflow averages // KeyName: worker.ESAnalyzerTimeWindow // Value type: Duration // Default value: 30 days ESAnalyzerTimeWindow // ESAnalyzerBufferWaitTime controls min time required to consider a worklow stuck // KeyName: worker.ESAnalyzerBufferWaitTime // Value type: Duration // Default value: 30 minutes ESAnalyzerBufferWaitTime // IsolationGroupStateRefreshInterval // KeyName: system.isolationGroupStateRefreshInterval // Value type: Duration // Default value: 30 seconds IsolationGroupStateRefreshInterval // IsolationGroupStateFetchTimeout is the dynamic config DB fetch timeout value // KeyName: system.isolationGroupStateFetchTimeout // Value type: Duration // Default value: 30 seconds IsolationGroupStateFetchTimeout // IsolationGroupStateUpdateTimeout is the dynamic config DB update timeout value // KeyName: system.isolationGroupStateUpdateTimeout // Value type: Duration // Default value: 30 seconds IsolationGroupStateUpdateTimeout // AsyncTaskDispatchTimeout is the timeout of dispatching tasks for async match // KeyName: matching.asyncTaskDispatchTimeout // Value type: Duration // Default value: 3 seconds // Allowed filters: domainName, taskListName, taskListType AsyncTaskDispatchTimeout // HistoryGlobalRatelimiterDecayAfter defines how long to wait for an update before considering a host's data "possibly gone", causing its weight to gradually decline. // KeyName: history.globalRatelimiterDecayAfter // Value type: Duration // Default value: 6 seconds HistoryGlobalRatelimiterDecayAfter // HistoryGlobalRatelimiterGCAfter defines how long to wait until a host's data is considered entirely useless, e.g. host has likely disappeared, its weight is very low, and the data can be deleted. // KeyName: history.globalRatelimiterGCAfter // Value type: Duration // Default value: 30 seconds HistoryGlobalRatelimiterGCAfter // LocalPollWaitTime is the wait time for a poller to wait before considering request forwarding // KeyName: matching.localPollWaitTime // Value type: Duration // Default value: 10ms // Allowed filters: domainName, taskListName, taskListType LocalPollWaitTime // LocalTaskWaitTime is the wait time for a task to wait before considering task forwarding // KeyName: matching.localTaskWaitTime // Value type: Duration // Default value: 10ms // Allowed filters: domainName, taskListName, taskListType LocalTaskWaitTime // LastDurationKey must be the last one in this const group LastDurationKey )
func (DurationKey) DefaultDuration ¶ added in v0.25.0
func (k DurationKey) DefaultDuration() time.Duration
func (DurationKey) DefaultValue ¶ added in v0.25.0
func (k DurationKey) DefaultValue() interface{}
func (DurationKey) Description ¶ added in v0.25.0
func (k DurationKey) Description() string
func (DurationKey) Filters ¶ added in v1.2.1
func (k DurationKey) Filters() []Filter
func (DurationKey) String ¶ added in v0.25.0
func (k DurationKey) String() string
type DurationPropertyFn ¶
type DurationPropertyFn func(opts ...FilterOption) time.Duration
DurationPropertyFn is a wrapper to get duration property from dynamic config
type DurationPropertyFnWithDomainFilter ¶
DurationPropertyFnWithDomainFilter is a wrapper to get duration property from dynamic config with domain as filter
type DurationPropertyFnWithDomainIDFilter ¶
DurationPropertyFnWithDomainIDFilter is a wrapper to get duration property from dynamic config with domainID as filter
type DurationPropertyFnWithShardIDFilter ¶
DurationPropertyFnWithShardIDFilter is a wrapper to get duration property from dynamic config with shardID as filter
type DurationPropertyFnWithTaskListInfoFilters ¶
type DurationPropertyFnWithTaskListInfoFilters func(domain string, taskList string, taskType int) time.Duration
DurationPropertyFnWithTaskListInfoFilters is a wrapper to get duration property from dynamic config with three filters: domain, taskList, taskType
type DurationPropertyFnWithWorkflowTypeFilter ¶ added in v0.24.0
type DurationPropertyFnWithWorkflowTypeFilter func(domainName string, workflowType string) time.Duration
DurationPropertyFnWithDomainFilter is a wrapper to get duration property from dynamic config with domain as filter
type DynamicBool ¶ added in v0.25.0
type DynamicDuration ¶ added in v0.25.0
type DynamicFloat ¶ added in v0.25.0
type DynamicInt ¶ added in v0.25.0
DynamicInt defines the properties for a dynamic config with int value type
type DynamicList ¶ added in v1.0.0
type DynamicMap ¶ added in v0.25.0
type DynamicString ¶ added in v0.25.0
type FileBasedClientConfig ¶
type FileBasedClientConfig struct { Filepath string `yaml:"filepath"` PollInterval time.Duration `yaml:"pollInterval"` }
FileBasedClientConfig is the config for the file based dynamic config client. It specifies where the config file is stored and how often the config should be updated by checking the config file again.
type Filter ¶
type Filter int
Filter represents a filter on the dynamic config key
const ( UnknownFilter Filter = iota // DomainName is the domain name DomainName // DomainID is the domain id DomainID // TaskListName is the tasklist name TaskListName // TaskType is the task type (0:Decision, 1:Activity) TaskType // ShardID is the shard id ShardID // ClusterName is the cluster name in a multi-region setup ClusterName // WorkflowID is the workflow id WorkflowID // WorkflowType is the workflow type name WorkflowType // RatelimitKey is the global ratelimit key (not a local key name) RatelimitKey // LastFilterTypeForTest must be the last one in this const group for testing purpose LastFilterTypeForTest )
func ParseFilter ¶ added in v0.23.1
type FilterOption ¶
type FilterOption func(filterMap map[Filter]interface{})
FilterOption is used to provide filters for dynamic config keys
func ClusterNameFilter ¶
func ClusterNameFilter(clusterName string) FilterOption
ClusterNameFilter filters by cluster name
func DomainIDFilter ¶
func DomainIDFilter(domainID string) FilterOption
DomainIDFilter filters by domain id
func RatelimitKeyFilter ¶ added in v1.2.12
func RatelimitKeyFilter(key string) FilterOption
RatelimitKeyFilter filters on global ratelimiter keys (via the global name, not local names)
func TaskListFilter ¶
func TaskListFilter(name string) FilterOption
TaskListFilter filters by task list name
func TaskTypeFilter ¶
func TaskTypeFilter(taskType int) FilterOption
TaskTypeFilter filters by task type
func WorkflowIDFilter ¶ added in v0.24.0
func WorkflowIDFilter(workflowID string) FilterOption
WorkflowIDFilter filters by workflowID
func WorkflowTypeFilter ¶ added in v0.24.0
func WorkflowTypeFilter(name string) FilterOption
WorkflowType filters by workflow type name
type FloatKey ¶ added in v0.25.0
type FloatKey int
const ( UnknownFloatKey FloatKey = iota // key for tests TestGetFloat64PropertyKey TestGetFloat64PropertyFilteredByShardIDKey // PersistenceErrorInjectionRate is rate for injecting random error in persistence // KeyName: system.persistenceErrorInjectionRate // Value type: Float64 // Default value: 0 // Allowed filters: N/A PersistenceErrorInjectionRate // AdminErrorInjectionRate is the rate for injecting random error in admin client // KeyName: admin.errorInjectionRate // Value type: Float64 // Default value: 0 // Allowed filters: N/A AdminErrorInjectionRate // DomainFailoverRefreshTimerJitterCoefficient is the jitter for domain failover refresh timer jitter // KeyName: frontend.domainFailoverRefreshTimerJitterCoefficient // Value type: Float64 // Default value: 0.1 // Allowed filters: N/A DomainFailoverRefreshTimerJitterCoefficient // FrontendErrorInjectionRate is rate for injecting random error in frontend client // KeyName: frontend.errorInjectionRate // Value type: Float64 // Default value: 0 // Allowed filters: N/A FrontendErrorInjectionRate // MatchingErrorInjectionRate is rate for injecting random error in matching client // KeyName: matching.errorInjectionRate // Value type: Float64 // Default value: 0 // Allowed filters: N/A MatchingErrorInjectionRate // TaskRedispatchIntervalJitterCoefficient is the task redispatch interval jitter coefficient // KeyName: history.taskRedispatchIntervalJitterCoefficient // Value type: Float64 // Default value: 0.15 // Allowed filters: N/A TaskRedispatchIntervalJitterCoefficient // QueueProcessorRandomSplitProbability is the probability for a domain to be split to a new processing queue // KeyName: history.queueProcessorRandomSplitProbability // Value type: Float64 // Default value: 0.01 // Allowed filters: N/A QueueProcessorRandomSplitProbability // QueueProcessorPollBackoffIntervalJitterCoefficient is backoff interval jitter coefficient // KeyName: history.queueProcessorPollBackoffIntervalJitterCoefficient // Value type: Float64 // Default value: 0.15 // Allowed filters: N/A QueueProcessorPollBackoffIntervalJitterCoefficient // TimerProcessorUpdateAckIntervalJitterCoefficient is the update interval jitter coefficient // KeyName: history.timerProcessorUpdateAckIntervalJitterCoefficient // Value type: Float64 // Default value: 0.15 // Allowed filters: N/A TimerProcessorUpdateAckIntervalJitterCoefficient // TimerProcessorMaxPollIntervalJitterCoefficient is the max poll interval jitter coefficient // KeyName: history.timerProcessorMaxPollIntervalJitterCoefficient // Value type: Float64 // Default value: 0.15 // Allowed filters: N/A TimerProcessorMaxPollIntervalJitterCoefficient // TimerProcessorSplitQueueIntervalJitterCoefficient is the split processing queue interval jitter coefficient // KeyName: history.timerProcessorSplitQueueIntervalJitterCoefficient // Value type: Float64 // Default value: 0.15 // Allowed filters: N/A TimerProcessorSplitQueueIntervalJitterCoefficient // TransferProcessorMaxPollIntervalJitterCoefficient is the max poll interval jitter coefficient // KeyName: history.transferProcessorMaxPollIntervalJitterCoefficient // Value type: Float64 // Default value: 0.15 // Allowed filters: N/A TransferProcessorMaxPollIntervalJitterCoefficient // TransferProcessorSplitQueueIntervalJitterCoefficient is the split processing queue interval jitter coefficient // KeyName: history.transferProcessorSplitQueueIntervalJitterCoefficient // Value type: Float64 // Default value: 0.15 // Allowed filters: N/A TransferProcessorSplitQueueIntervalJitterCoefficient // TransferProcessorUpdateAckIntervalJitterCoefficient is the update interval jitter coefficient // KeyName: history.transferProcessorUpdateAckIntervalJitterCoefficient // Value type: Float64 // Default value: 0.15 // Allowed filters: N/A TransferProcessorUpdateAckIntervalJitterCoefficient // ReplicationTaskProcessorCleanupJitterCoefficient is the jitter for cleanup timer // KeyName: history.ReplicationTaskProcessorCleanupJitterCoefficient // Value type: Float64 // Default value: 0.15 // Allowed filters: ShardID ReplicationTaskProcessorCleanupJitterCoefficient // ReplicationTaskProcessorStartWaitJitterCoefficient is the jitter for batch start wait timer // KeyName: history.ReplicationTaskProcessorStartWaitJitterCoefficient // Value type: Float64 // Default value: 0.9 // Allowed filters: ShardID ReplicationTaskProcessorStartWaitJitterCoefficient // ReplicationTaskProcessorHostQPS is the qps of task processing rate limiter on host level // KeyName: history.ReplicationTaskProcessorHostQPS // Value type: Float64 // Default value: 1500 // Allowed filters: N/A ReplicationTaskProcessorHostQPS // ReplicationTaskProcessorShardQPS is the qps of task processing rate limiter on shard level // KeyName: history.ReplicationTaskProcessorShardQPS // Value type: Float64 // Default value: 5 // Allowed filters: N/A ReplicationTaskProcessorShardQPS // ReplicationTaskGenerationQPS is the wait time between each replication task generation qps // KeyName: history.ReplicationTaskGenerationQPS // Value type: Float64 // Default value: 100 // Allowed filters: N/A ReplicationTaskGenerationQPS // MutableStateChecksumInvalidateBefore is the epoch timestamp before which all checksums are to be discarded // KeyName: history.mutableStateChecksumInvalidateBefore // Value type: Float64 // Default value: 0 // Allowed filters: N/A MutableStateChecksumInvalidateBefore // NotifyFailoverMarkerTimerJitterCoefficient is the jitter for failover marker notifier timer // KeyName: history.NotifyFailoverMarkerTimerJitterCoefficient // Value type: Float64 // Default value: 0.15 // Allowed filters: N/A NotifyFailoverMarkerTimerJitterCoefficient // HistoryErrorInjectionRate is rate for injecting random error in history client // KeyName: history.errorInjectionRate // Value type: Float64 // Default value: 0 // Allowed filters: N/A HistoryErrorInjectionRate // ReplicationTaskFetcherTimerJitterCoefficient is the jitter for fetcher timer // KeyName: history.ReplicationTaskFetcherTimerJitterCoefficient // Value type: Float64 // Default value: 0.15 // Allowed filters: N/A ReplicationTaskFetcherTimerJitterCoefficient // WorkerDeterministicConstructionCheckProbability controls the probability of running a deterministic construction check for any given archival // KeyName: N/A // Default value: N/A // TODO: https://github.com/uber/cadence/issues/3861 WorkerDeterministicConstructionCheckProbability // WorkerBlobIntegrityCheckProbability controls the probability of running an integrity check for any given archival // KeyName: N/A // Default value: N/A // TODO: https://github.com/uber/cadence/issues/3861 WorkerBlobIntegrityCheckProbability // HistoryGlobalRatelimiterNewDataWeight defines how much weight to give each host's newest data, per update. Must be between 0 and 1, higher values match new values more closely after a single update. // KeyName: history.globalRatelimiterNewDataWeight // Value type: Float64 // Default value: 0.5 HistoryGlobalRatelimiterNewDataWeight // LastFloatKey must be the last one in this const group LastFloatKey )
func (FloatKey) DefaultFloat ¶ added in v0.25.0
func (FloatKey) DefaultValue ¶ added in v0.25.0
func (k FloatKey) DefaultValue() interface{}
func (FloatKey) Description ¶ added in v0.25.0
type FloatPropertyFn ¶
type FloatPropertyFn func(opts ...FilterOption) float64
FloatPropertyFn is a wrapper to get float property from dynamic config
func (FloatPropertyFn) AsFloat64 ¶ added in v0.24.0
func (f FloatPropertyFn) AsFloat64(opts ...FilterOption) func() float64
type FloatPropertyFnWithShardIDFilter ¶
FloatPropertyFnWithShardIDFilter is a wrapper to get float property from dynamic config with shardID as filter
type IntKey ¶ added in v0.25.0
type IntKey int
const ( UnknownIntKey IntKey = iota // key for tests TestGetIntPropertyKey TestGetIntPropertyFilteredByDomainKey TestGetIntPropertyFilteredByWorkflowTypeKey TestGetIntPropertyFilteredByTaskListInfoKey TestGetIntPropertyFilteredByShardIDKey TransactionSizeLimit MaxRetentionDays MinRetentionDays MaxDecisionStartToCloseSeconds BlobSizeLimitError // BlobSizeLimitWarn is the per event blob size limit for warning // KeyName: limit.blobSize.warn // Value type: Int // Default value: 262144 (256*1024) // Allowed filters: DomainName BlobSizeLimitWarn // HistorySizeLimitError is the per workflow execution history size limit // KeyName: limit.historySize.error // Value type: Int // Default value: 209715200 (200*1024*1024) // Allowed filters: DomainName HistorySizeLimitError // HistorySizeLimitWarn is the per workflow execution history size limit for warning // KeyName: limit.historySize.warn // Value type: Int // Default value: 52428800 (50*1024*1024) // Allowed filters: DomainName HistorySizeLimitWarn // HistoryCountLimitError is the per workflow execution history event count limit // KeyName: limit.historyCount.error // Value type: Int // Default value: 204800 (200*1024) // Allowed filters: DomainName HistoryCountLimitError // HistoryCountLimitWarn is the per workflow execution history event count limit for warning // KeyName: limit.historyCount.warn // Value type: Int // Default value: 51200 (50*1024) // Allowed filters: DomainName HistoryCountLimitWarn // PendingActivitiesCountLimitError is the limit of how many pending activities a workflow can have at a point in time // KeyName: limit.pendingActivityCount.error // Value type: Int // Default value: 1024 PendingActivitiesCountLimitError // PendingActivitiesCountLimitWarn is the limit of how many activities a workflow can have before a warning is logged // KeyName: limit.pendingActivityCount.warn // Value type: Int // Default value: 512 PendingActivitiesCountLimitWarn // DomainNameMaxLength is the length limit for domain name // KeyName: limit.domainNameLength // Value type: Int // Default value: common.DefaultIDLengthErrorLimit (1000) // Allowed filters: DomainName DomainNameMaxLength // IdentityMaxLength is the length limit for identity // KeyName: limit.identityLength // Value type: Int // Default value: 1000 ( see common.DefaultIDLengthErrorLimit) // Allowed filters: DomainName IdentityMaxLength // WorkflowIDMaxLength is the length limit for workflowID // KeyName: limit.workflowIDLength // Value type: Int // Default value: 1000 (see common.DefaultIDLengthErrorLimit) // Allowed filters: DomainName WorkflowIDMaxLength // SignalNameMaxLength is the length limit for signal name // KeyName: limit.signalNameLength // Value type: Int // Default value: 1000 (see common.DefaultIDLengthErrorLimit) // Allowed filters: DomainName SignalNameMaxLength // WorkflowTypeMaxLength is the length limit for workflow type // KeyName: limit.workflowTypeLength // Value type: Int // Default value: 1000 (see common.DefaultIDLengthErrorLimit) // Allowed filters: DomainName WorkflowTypeMaxLength // RequestIDMaxLength is the length limit for requestID // KeyName: limit.requestIDLength // Value type: Int // Default value: 1000 (see common.DefaultIDLengthErrorLimit) // Allowed filters: DomainName RequestIDMaxLength // TaskListNameMaxLength is the length limit for task list name // KeyName: limit.taskListNameLength // Value type: Int // Default value: 1000 (see common.DefaultIDLengthErrorLimit) // Allowed filters: DomainName TaskListNameMaxLength // ActivityIDMaxLength is the length limit for activityID // KeyName: limit.activityIDLength // Value type: Int // Default value: 1000 (see common.DefaultIDLengthErrorLimit) // Allowed filters: DomainName ActivityIDMaxLength // ActivityTypeMaxLength is the length limit for activity type // KeyName: limit.activityTypeLength // Value type: Int // Default value: 1000 (see common.DefaultIDLengthErrorLimit) // Allowed filters: DomainName ActivityTypeMaxLength // MarkerNameMaxLength is the length limit for marker name // KeyName: limit.markerNameLength // Value type: Int // Default value: 1000 (see common.DefaultIDLengthErrorLimit) // Allowed filters: DomainName MarkerNameMaxLength // TimerIDMaxLength is the length limit for timerID // KeyName: limit.timerIDLength // Value type: Int // Default value: 1000 (see common.DefaultIDLengthErrorLimit) // Allowed filters: DomainName TimerIDMaxLength // MaxIDLengthWarnLimit is the warn length limit for various IDs, including: Domain, TaskList, WorkflowID, ActivityID, TimerID, WorkflowType, ActivityType, SignalName, MarkerName, ErrorReason/FailureReason/CancelCause, Identity, RequestID // KeyName: limit.maxIDWarnLength // Value type: Int // Default value: 128 (see common.DefaultIDLengthWarnLimit) // Allowed filters: N/A MaxIDLengthWarnLimit // FrontendPersistenceMaxQPS is the max qps frontend host can query DB // KeyName: frontend.persistenceMaxQPS // Value type: Int // Default value: 2000 // Allowed filters: N/A FrontendPersistenceMaxQPS // FrontendPersistenceGlobalMaxQPS is the max qps frontend cluster can query DB // KeyName: frontend.persistenceGlobalMaxQPS // Value type: Int // Default value: 0 // Allowed filters: N/A FrontendPersistenceGlobalMaxQPS // FrontendVisibilityMaxPageSize is default max size for ListWorkflowExecutions in one page // KeyName: frontend.visibilityMaxPageSize // Value type: Int // Default value: 1000 // Allowed filters: DomainName FrontendVisibilityMaxPageSize // FrontendVisibilityListMaxQPS is max qps frontend can list open/close workflows // KeyName: frontend.visibilityListMaxQPS // Value type: Int // Default value: 10 // Allowed filters: DomainName // deprecated: never used for ratelimiting, only sampling-based failure injection, and only on database-based visibility FrontendVisibilityListMaxQPS // FrontendESVisibilityListMaxQPS is max qps frontend can list open/close workflows from ElasticSearch // KeyName: frontend.esVisibilityListMaxQPS // Value type: Int // Default value: 30 // Allowed filters: DomainName // deprecated: never read from, all ES reads and writes erroneously use PersistenceMaxQPS FrontendESVisibilityListMaxQPS // FrontendESIndexMaxResultWindow is ElasticSearch index setting max_result_window // KeyName: frontend.esIndexMaxResultWindow // Value type: Int // Default value: 10000 // Allowed filters: N/A FrontendESIndexMaxResultWindow // FrontendHistoryMaxPageSize is default max size for GetWorkflowExecutionHistory in one page // KeyName: frontend.historyMaxPageSize // Value type: Int // Default value: 1000 (see common.GetHistoryMaxPageSize) // Allowed filters: DomainName FrontendHistoryMaxPageSize // FrontendUserRPS is used to limit "user" requests (StartWorkflow, Signal, etc) // per frontend instance (across all domains, or for non-domain-related requests), // and is mostly intended to protect against excessive single-host load. // // KeyName: frontend.rps // Value type: Int // Default value: 1200 // Allowed filters: N/A FrontendUserRPS // FrontendWorkerRPS is used to limit "worker" requests (PollFor...Task, RespondTask..., etc) // per frontend instance (across all domains, or for non-domain-related requests), // and is mostly intended to protect against excessive single-host load. // // KeyName: frontend.workerrps // Value type: Int // Default value: UnlimitedRPS // Allowed filters: N/A FrontendWorkerRPS // FrontendVisibilityRPS is used to limit "visibility" requests (ListWorkflow* and similar) // per frontend instance (across all domains, or for non-domain-related requests), // and is mostly intended to protect against excessive single-host load. // // KeyName: frontend.visibilityrps // Value type: Int // Default value: UnlimitedRPS // Allowed filters: N/A FrontendVisibilityRPS // FrontendAsyncRPS is used to limit "async" requests (StartWorkflowAsync, etc for many "user" APIs) // per frontend instance (across all domains, or for non-domain-related requests), // and is mostly intended to protect against excessive single-host load. // // KeyName: frontend.asyncrps // Value type: Int // Default value: 10000 // Allowed filters: N/A FrontendAsyncRPS // FrontendMaxDomainUserRPSPerInstance is used to limit "user" requests (StartWorkflow, Signal, etc) // per domain per frontend instance, and is mostly intended to protect against excessive single-host load. // // This limit applies along-side FrontendGlobalDomainUserRPS: both must be allowed to allow a request. // // KeyName: frontend.domainrps // Value type: Int // Default value: 1200 // Allowed filters: DomainName FrontendMaxDomainUserRPSPerInstance // FrontendMaxDomainWorkerRPSPerInstance is used to limit "worker" requests (PollFor...Task, RespondTask..., etc) // per domain per frontend instance, and is mostly intended to protect against excessive single-host load. // // This limit applies along-side FrontendGlobalDomainWorkerRPS: both must be allowed to allow a request. // // KeyName: frontend.domainworkerrps // Value type: Int // Default value: UnlimitedRPS // Allowed filters: DomainName FrontendMaxDomainWorkerRPSPerInstance // FrontendMaxDomainVisibilityRPSPerInstance is used to limit "visibility" requests (ListWorkflow* and similar) // per domain per frontend instance, and is mostly intended to protect against excessive single-host load. // // This limit applies along-side FrontendGlobalDomainVisibilityRPS: both must be allowed to allow a request. // // KeyName: frontend.domainvisibilityrps // Value type: Int // Default value: UnlimitedRPS // Allowed filters: DomainName FrontendMaxDomainVisibilityRPSPerInstance // FrontendMaxDomainAsyncRPSPerInstance is used to limit "async" requests (StartWorkflowAsync, etc for many "user" APIs) // per frontend instance, and is mostly intended to protect against excessive single-host load. // // This limit applies along-side FrontendGlobalDomainAsyncRPS: both must be allowed to allow a request. // // KeyName: frontend.domainasyncrps // Value type: Int // Default value: 10000 // Allowed filters: DomainName FrontendMaxDomainAsyncRPSPerInstance // FrontendGlobalDomainUserRPS is used to limit "user" requests (StartWorkflow, Signal, etc) // per domain to a target RPS that is shared across the entire cluster. // // Currently, there are two ways this is achieved, which can be selected by FrontendGlobalRatelimiterMode // with a "user:" key prefix: // 1. "local", where the configured RPS is split evenly across all frontend hosts in the cluster. // This works well if your load is roughly evenly distributed. // 2. "global", where frontend hosts share load information with each other, to adjust to imbalanced load. // This works well if your load is very imbalanced, e.g. one domain tends to contact a subset of frontend hosts much more than others. // // KeyName: frontend.globalDomainrps // Value type: Int // Default value: UnlimitedRPS (0 triggers a fallback to per-instance-RPS, generally avoid) // Allowed filters: DomainName FrontendGlobalDomainUserRPS // FrontendGlobalDomainWorkerRPS is used to limit "worker" requests (PollFor...Task, RespondTask..., etc) // per domain to a target RPS that is shared across the entire cluster. // // Currently, there are two ways this is achieved, which can be selected by FrontendGlobalRatelimiterMode // with a "worker:" key prefix: // 1. "local", where the configured RPS is split evenly across all frontend hosts in the cluster. // This works well if your load is roughly evenly distributed. // 2. "global", where frontend hosts share load information with each other, to adjust to imbalanced load. // This works well if your load is very imbalanced, e.g. one domain tends to contact a subset of frontend hosts much more than others. // // KeyName: frontend.globalDomainWorkerrps // Value type: Int // Default value: UnlimitedRPS (0 triggers a fallback to per-instance-RPS, generally avoid) // Allowed filters: DomainName FrontendGlobalDomainWorkerRPS // FrontendGlobalDomainVisibilityRPS is used to limit "visibility" requests (ListWorkflow* and similar) // per domain to a target RPS that is shared across the entire cluster. // // Currently, there are two ways this is achieved, which can be selected by FrontendGlobalRatelimiterMode // with a "visibility:" key prefix: // 1. "local", where the configured RPS is split evenly across all frontend hosts in the cluster. // This works well if your load is roughly evenly distributed. // 2. "global", where frontend hosts share load information with each other, to adjust to imbalanced load. // This works well if your load is very imbalanced, e.g. one domain tends to contact a subset of frontend hosts much more than others. // // KeyName: frontend.globalDomainVisibilityrps // Value type: Int // Default value: UnlimitedRPS (0 triggers a fallback to per-instance-RPS, generally avoid) // Allowed filters: DomainName FrontendGlobalDomainVisibilityRPS // FrontendGlobalDomainAsyncRPS is used to limit "async" requests (StartWorkflowAsync, etc for many "user" APIs) // per domain to a target RPS that is shared across the entire cluster. // // Currently, there are two ways this is achieved, which can be selected by FrontendGlobalRatelimiterMode // with a "async:" key prefix: // 1. "local", where the configured RPS is split evenly across all frontend hosts in the cluster. // This works well if your load is roughly evenly distributed. // 2. "global", where frontend hosts share load information with each other, to adjust to imbalanced load. // This works well if your load is very imbalanced, e.g. one domain tends to contact a subset of frontend hosts much more than others. // // KeyName: frontend.globalDomainAsyncrps // Value type: Int // Default value: 100000 // Allowed filters: DomainName FrontendGlobalDomainAsyncRPS // FrontendDecisionResultCountLimit is max number of decisions per RespondDecisionTaskCompleted request // KeyName: frontend.decisionResultCountLimit // Value type: Int // Default value: 0 // Allowed filters: DomainName FrontendDecisionResultCountLimit // FrontendHistoryMgrNumConns is for persistence cluster.NumConns // KeyName: frontend.historyMgrNumConns // Value type: Int // Default value: 10 // Allowed filters: N/A // Deprecated: not used FrontendHistoryMgrNumConns // FrontendThrottledLogRPS is the rate limit on number of log messages emitted per second for throttled logger // KeyName: frontend.throttledLogRPS // Value type: Int // Default value: 20 // Allowed filters: N/A FrontendThrottledLogRPS // FrontendMaxBadBinaries is the max number of bad binaries in domain config // KeyName: frontend.maxBadBinaries // Value type: Int // Default value: 10 (see domain.MaxBadBinaries) // Allowed filters: DomainName FrontendMaxBadBinaries // SearchAttributesNumberOfKeysLimit is the limit of number of keys // KeyName: frontend.searchAttributesNumberOfKeysLimit // Value type: Int // Default value: 100 // Allowed filters: DomainName SearchAttributesNumberOfKeysLimit // SearchAttributesSizeOfValueLimit is the size limit of each value // KeyName: frontend.searchAttributesSizeOfValueLimit // Value type: Int // Default value: 2048 (2*1024) // Allowed filters: DomainName SearchAttributesSizeOfValueLimit // SearchAttributesTotalSizeLimit is the size limit of the whole map // KeyName: frontend.searchAttributesTotalSizeLimit // Value type: Int // Default value: 40960 (40*1024) // Allowed filters: DomainName SearchAttributesTotalSizeLimit // VisibilityArchivalQueryMaxPageSize is the maximum page size for a visibility archival query // KeyName: frontend.visibilityArchivalQueryMaxPageSize // Value type: Int // Default value: 10000 // Allowed filters: N/A VisibilityArchivalQueryMaxPageSize // FrontendFailoverHistoryMaxSize is the maximum size for the number of failover event records in a domain failover history // KeyName: frontend.failoverHistoryMaxSize // Value type: Int // Default value: 5 // Allowed filters: DomainName FrontendFailoverHistoryMaxSize // MatchingUserRPS is request rate per second for each matching host // KeyName: matching.rps // Value type: Int // Default value: 1200 // Allowed filters: N/A MatchingUserRPS // MatchingWorkerRPS is background-processing request rate per second for each matching host // KeyName: matching.workerrps // Value type: Int // Default value: UnlimitedRPS // Allowed filters: N/A MatchingWorkerRPS // MatchingDomainUserRPS is request rate per domain per second for each matching host // KeyName: matching.domainrps // Value type: Int // Default value: 0 // Allowed filters: N/A MatchingDomainUserRPS // MatchingDomainWorkerRPS is background-processing request rate per domain per second for each matching host // KeyName: matching.domainworkerrps // Value type: Int // Default value: UnlimitedRPS // Allowed filters: N/A MatchingDomainWorkerRPS // MatchingPersistenceMaxQPS is the max qps matching host can query DB // KeyName: matching.persistenceMaxQPS // Value type: Int // Default value: 3000 // Allowed filters: N/A MatchingPersistenceMaxQPS // MatchingPersistenceGlobalMaxQPS is the max qps matching cluster can query DB // KeyName: matching.persistenceGlobalMaxQPS // Value type: Int // Default value: 0 // Allowed filters: N/A MatchingPersistenceGlobalMaxQPS // MatchingMinTaskThrottlingBurstSize is the minimum burst size for task list throttling // KeyName: matching.minTaskThrottlingBurstSize // Value type: Int // Default value: 1 // Allowed filters: DomainName,TasklistName,TasklistType MatchingMinTaskThrottlingBurstSize // MatchingGetTasksBatchSize is the maximum batch size to fetch from the task buffer // KeyName: matching.getTasksBatchSize // Value type: Int // Default value: 1000 // Allowed filters: DomainName,TasklistName,TasklistType MatchingGetTasksBatchSize // MatchingOutstandingTaskAppendsThreshold is the threshold for outstanding task appends // KeyName: matching.outstandingTaskAppendsThreshold // Value type: Int // Default value: 250 // Allowed filters: DomainName,TasklistName,TasklistType MatchingOutstandingTaskAppendsThreshold // MatchingMaxTaskBatchSize is max batch size for task writer // KeyName: matching.maxTaskBatchSize // Value type: Int // Default value: 100 // Allowed filters: DomainName,TasklistName,TasklistType MatchingMaxTaskBatchSize // MatchingMaxTaskDeleteBatchSize is the max batch size for range deletion of tasks // KeyName: matching.maxTaskDeleteBatchSize // Value type: Int // Default value: 100 // Allowed filters: DomainName,TasklistName,TasklistType MatchingMaxTaskDeleteBatchSize // MatchingThrottledLogRPS is the rate limit on number of log messages emitted per second for throttled logger // KeyName: matching.throttledLogRPS // Value type: Int // Default value: 20 // Allowed filters: N/A MatchingThrottledLogRPS // MatchingNumTasklistWritePartitions is the number of write partitions for a task list // KeyName: matching.numTasklistWritePartitions // Value type: Int // Default value: 1 // Allowed filters: DomainName,TasklistName,TasklistType MatchingNumTasklistWritePartitions // MatchingNumTasklistReadPartitions is the number of read partitions for a task list // KeyName: matching.numTasklistReadPartitions // Value type: Int // Default value: 1 // Allowed filters: DomainName,TasklistName,TasklistType MatchingNumTasklistReadPartitions // MatchingForwarderMaxOutstandingPolls is the max number of inflight polls from the forwarder // KeyName: matching.forwarderMaxOutstandingPolls // Value type: Int // Default value: 1 // Allowed filters: DomainName,TasklistName,TasklistType MatchingForwarderMaxOutstandingPolls // MatchingForwarderMaxOutstandingTasks is the max number of inflight addTask/queryTask from the forwarder // KeyName: matching.forwarderMaxOutstandingTasks // Value type: Int // Default value: 1 // Allowed filters: DomainName,TasklistName,TasklistType MatchingForwarderMaxOutstandingTasks // MatchingForwarderMaxRatePerSecond is the max rate at which add/query can be forwarded // KeyName: matching.forwarderMaxRatePerSecond // Value type: Int // Default value: 10 // Allowed filters: DomainName,TasklistName,TasklistType MatchingForwarderMaxRatePerSecond // MatchingForwarderMaxChildrenPerNode is the max number of children per node in the task list partition tree // KeyName: matching.forwarderMaxChildrenPerNode // Value type: Int // Default value: 20 // Allowed filters: DomainName,TasklistName,TasklistType MatchingForwarderMaxChildrenPerNode // HistoryRPS is request rate per second for each history host // KeyName: history.rps // Value type: Int // Default value: 3000 // Allowed filters: N/A HistoryRPS // HistoryPersistenceMaxQPS is the max qps history host can query DB // KeyName: history.persistenceMaxQPS // Value type: Int // Default value: 9000 // Allowed filters: N/A HistoryPersistenceMaxQPS // HistoryPersistenceGlobalMaxQPS is the max qps history cluster can query DB // KeyName: history.persistenceGlobalMaxQPS // Value type: Int // Default value: 0 // Allowed filters: N/A HistoryPersistenceGlobalMaxQPS // HistoryVisibilityOpenMaxQPS is max qps one history host can write visibility open_executions // KeyName: history.historyVisibilityOpenMaxQPS // Value type: Int // Default value: 300 // Allowed filters: DomainName HistoryVisibilityOpenMaxQPS // HistoryVisibilityClosedMaxQPS is max qps one history host can write visibility closed_executions // KeyName: history.historyVisibilityClosedMaxQPS // Value type: Int // Default value: 300 // Allowed filters: DomainName HistoryVisibilityClosedMaxQPS // HistoryCacheInitialSize is initial size of history cache // KeyName: history.cacheInitialSize // Value type: Int // Default value: 128 // Allowed filters: N/A HistoryCacheInitialSize // HistoryCacheMaxSize is max size of history cache // KeyName: history.cacheMaxSize // Value type: Int // Default value: 512 // Allowed filters: N/A HistoryCacheMaxSize // EventsCacheInitialCount is initial count of events cache // KeyName: history.eventsCacheInitialSize // Value type: Int // Default value: 128 // Allowed filters: N/A EventsCacheInitialCount // EventsCacheMaxCount is max count of events cache // KeyName: history.eventsCacheMaxSize // Value type: Int // Default value: 512 // Allowed filters: N/A EventsCacheMaxCount // EventsCacheMaxSize is max size of events cache in bytes // KeyName: history.eventsCacheMaxSizeInBytes // Value type: Int // Default value: 0 // Allowed filters: N/A EventsCacheMaxSize // EventsCacheGlobalInitialCount is initial count of global events cache // KeyName: history.eventsCacheGlobalInitialSize // Value type: Int // Default value: 4096 // Allowed filters: N/A EventsCacheGlobalInitialCount // EventsCacheGlobalMaxCount is max count of global events cache // KeyName: history.eventsCacheGlobalMaxSize // Value type: Int // Default value: 131072 // Allowed filters: N/A EventsCacheGlobalMaxCount // AcquireShardConcurrency is number of goroutines that can be used to acquire shards in the shard controller. // KeyName: history.acquireShardConcurrency // Value type: Int // Default value: 1 // Allowed filters: N/A AcquireShardConcurrency // TaskProcessRPS is the task processing rate per second for each domain // KeyName: history.taskProcessRPS // Value type: Int // Default value: 1000 // Allowed filters: DomainName TaskProcessRPS // TaskSchedulerType is the task scheduler type for priority task processor // KeyName: history.taskSchedulerType // Value type: Int enum(1 for SchedulerTypeFIFO, 2 for SchedulerTypeWRR(weighted round robin scheduler implementation)) // Default value: 2 (task.SchedulerTypeWRR) // Allowed filters: N/A TaskSchedulerType // TaskSchedulerWorkerCount is the number of workers per host in task scheduler // KeyName: history.taskSchedulerWorkerCount // Value type: Int // Default value: 200 // Allowed filters: N/A TaskSchedulerWorkerCount // TaskSchedulerShardWorkerCount is the number of worker per shard in task scheduler // KeyName: history.taskSchedulerShardWorkerCount // Value type: Int // Default value: 0 // Allowed filters: N/A TaskSchedulerShardWorkerCount // TaskSchedulerQueueSize is the size of task channel for host level task scheduler // KeyName: history.taskSchedulerQueueSize // Value type: Int // Default value: 10000 // Allowed filters: N/A TaskSchedulerQueueSize // TaskSchedulerShardQueueSize is the size of task channel for shard level task scheduler // KeyName: history.taskSchedulerShardQueueSize // Value type: Int // Default value: 200 // Allowed filters: N/A TaskSchedulerShardQueueSize // TaskSchedulerDispatcherCount is the number of task dispatcher in task scheduler (only applies to host level task scheduler) // KeyName: history.taskSchedulerDispatcherCount // Value type: Int // Default value: 1 // Allowed filters: N/A TaskSchedulerDispatcherCount // TaskCriticalRetryCount is the critical retry count for background tasks // when task attempt exceeds this threshold: // - task attempt metrics and additional error logs will be emitted // - task priority will be lowered // KeyName: history.taskCriticalRetryCount // Value type: Int // Default value: 50 // Allowed filters: N/A TaskCriticalRetryCount // QueueProcessorSplitMaxLevel is the max processing queue level // KeyName: history.queueProcessorSplitMaxLevel // Value type: Int // Default value: 2 // 3 levels, start from 0 // Allowed filters: N/A QueueProcessorSplitMaxLevel // TimerTaskBatchSize is batch size for timer processor to process tasks // KeyName: history.timerTaskBatchSize // Value type: Int // Default value: 100 // Allowed filters: N/A TimerTaskBatchSize // TimerTaskDeleteBatchSize is batch size for timer processor to delete timer tasks // KeyName: history.timerTaskDeleteBatchSize // Value type: Int // Default value: 4000 // Allowed filters: N/A TimerTaskDeleteBatchSize // TimerProcessorGetFailureRetryCount is retry count for timer processor get failure operation // KeyName: history.timerProcessorGetFailureRetryCount // Value type: Int // Default value: 5 // Allowed filters: N/A TimerProcessorGetFailureRetryCount // TimerProcessorCompleteTimerFailureRetryCount is retry count for timer processor complete timer operation // KeyName: history.timerProcessorCompleteTimerFailureRetryCount // Value type: Int // Default value: 10 // Allowed filters: N/A TimerProcessorCompleteTimerFailureRetryCount // TimerProcessorFailoverMaxPollRPS is max poll rate per second for timer processor // KeyName: history.timerProcessorFailoverMaxPollRPS // Value type: Int // Default value: 1 // Allowed filters: N/A TimerProcessorFailoverMaxPollRPS // TimerProcessorMaxPollRPS is max poll rate per second for timer processor // KeyName: history.timerProcessorMaxPollRPS // Value type: Int // Default value: 20 // Allowed filters: N/A TimerProcessorMaxPollRPS // TimerProcessorMaxRedispatchQueueSize is the threshold of the number of tasks in the redispatch queue for timer processor // KeyName: history.timerProcessorMaxRedispatchQueueSize // Value type: Int // Default value: 10000 // Allowed filters: N/A TimerProcessorMaxRedispatchQueueSize // TimerProcessorHistoryArchivalSizeLimit is the max history size for inline archival // KeyName: history.timerProcessorHistoryArchivalSizeLimit // Value type: Int // Default value: 500*1024 // Allowed filters: N/A TimerProcessorHistoryArchivalSizeLimit // TransferTaskBatchSize is batch size for transferQueueProcessor // KeyName: history.transferTaskBatchSize // Value type: Int // Default value: 100 // Allowed filters: N/A TransferTaskBatchSize // TransferTaskDeleteBatchSize is batch size for transferQueueProcessor to delete transfer tasks // KeyName: history.transferTaskDeleteBatchSize // Value type: Int // Default value: 4000 // Allowed filters: N/A TransferTaskDeleteBatchSize // TransferProcessorFailoverMaxPollRPS is max poll rate per second for transferQueueProcessor // KeyName: history.transferProcessorFailoverMaxPollRPS // Value type: Int // Default value: 1 // Allowed filters: N/A TransferProcessorFailoverMaxPollRPS // TransferProcessorMaxPollRPS is max poll rate per second for transferQueueProcessor // KeyName: history.transferProcessorMaxPollRPS // Value type: Int // Default value: 20 // Allowed filters: N/A TransferProcessorMaxPollRPS // TransferProcessorCompleteTransferFailureRetryCount is times of retry for failure // KeyName: history.transferProcessorCompleteTransferFailureRetryCount // Value type: Int // Default value: 10 // Allowed filters: N/A TransferProcessorCompleteTransferFailureRetryCount // TransferProcessorMaxRedispatchQueueSize is the threshold of the number of tasks in the redispatch queue for transferQueueProcessor // KeyName: history.transferProcessorMaxRedispatchQueueSize // Value type: Int // Default value: 10000 // Allowed filters: N/A TransferProcessorMaxRedispatchQueueSize // ReplicatorTaskBatchSize is batch size for ReplicatorProcessor // KeyName: history.replicatorTaskBatchSize // Value type: Int // Default value: 25 // Allowed filters: N/A ReplicatorTaskBatchSize // ReplicatorTaskDeleteBatchSize is batch size for ReplicatorProcessor to delete replication tasks // KeyName: history.replicatorTaskDeleteBatchSize // Value type: Int // Default value: 4000 // Allowed filters: N/A ReplicatorTaskDeleteBatchSize // ReplicatorReadTaskMaxRetryCount is the number of read replication task retry time // KeyName: history.replicatorReadTaskMaxRetryCount // Value type: Int // Default value: 3 // Allowed filters: N/A ReplicatorReadTaskMaxRetryCount // ReplicatorCacheCapacity is the capacity of replication cache in number of tasks // KeyName: history.replicatorCacheCapacity // Value type: Int // Default value: 10000 // Allowed filters: N/A ReplicatorCacheCapacity // ExecutionMgrNumConns is persistence connections number for ExecutionManager // KeyName: history.executionMgrNumConns // Value type: Int // Default value: 50 // Allowed filters: N/A // Deprecated: not used ExecutionMgrNumConns // HistoryMgrNumConns is persistence connections number for HistoryManager // KeyName: history.historyMgrNumConns // Value type: Int // Default value: 50 // Allowed filters: N/A // Deprecated: not used HistoryMgrNumConns // MaximumBufferedEventsBatch is max number of buffer event in mutable state // KeyName: history.maximumBufferedEventsBatch // Value type: Int // Default value: 100 // Allowed filters: N/A MaximumBufferedEventsBatch // MaximumSignalsPerExecution is max number of signals supported by single execution // KeyName: history.maximumSignalsPerExecution // Value type: Int // Default value: 10000 // Allowed filters: DomainName MaximumSignalsPerExecution // NumArchiveSystemWorkflows is key for number of archive system workflows running in total // KeyName: history.numArchiveSystemWorkflows // Value type: Int // Default value: 1000 // Allowed filters: N/A NumArchiveSystemWorkflows // ArchiveRequestRPS is the rate limit on the number of archive request per second // KeyName: history.archiveRequestRPS // Value type: Int // Default value: 300 // should be much smaller than frontend RPS // Allowed filters: N/A ArchiveRequestRPS // ArchiveInlineHistoryRPS is the (per instance) rate limit on the number of inline history archival attempts per second // KeyName: history.archiveInlineHistoryRPS // Value type: Int // Default value: 1000 // Allowed filters: N/A ArchiveInlineHistoryRPS // ArchiveInlineHistoryGlobalRPS is the global rate limit on the number of inline history archival attempts per second // KeyName: history.archiveInlineHistoryGlobalRPS // Value type: Int // Default value: 10000 // Allowed filters: N/A ArchiveInlineHistoryGlobalRPS // ArchiveInlineVisibilityRPS is the (per instance) rate limit on the number of inline visibility archival attempts per second // KeyName: history.archiveInlineVisibilityRPS // Value type: Int // Default value: 1000 // Allowed filters: N/A ArchiveInlineVisibilityRPS // ArchiveInlineVisibilityGlobalRPS is the global rate limit on the number of inline visibility archival attempts per second // KeyName: history.archiveInlineVisibilityGlobalRPS // Value type: Int // Default value: 10000 // Allowed filters: N/A ArchiveInlineVisibilityGlobalRPS // HistoryMaxAutoResetPoints is the key for max number of auto reset points stored in mutableState // KeyName: history.historyMaxAutoResetPoints // Value type: Int // Default value: DefaultHistoryMaxAutoResetPoints // Allowed filters: DomainName HistoryMaxAutoResetPoints // ParentClosePolicyThreshold is decides that parent close policy will be processed by sys workers(if enabled) ifthe number of children greater than or equal to this threshold // KeyName: history.parentClosePolicyThreshold // Value type: Int // Default value: 10 // Allowed filters: DomainName ParentClosePolicyThreshold ParentClosePolicyBatchSize // NumParentClosePolicySystemWorkflows is key for number of parentClosePolicy system workflows running in total // KeyName: history.numParentClosePolicySystemWorkflows // Value type: Int // Default value: 10 // Allowed filters: N/A NumParentClosePolicySystemWorkflows // HistoryThrottledLogRPS is the rate limit on number of log messages emitted per second for throttled logger // KeyName: history.throttledLogRPS // Value type: Int // Default value: 4 // Allowed filters: N/A HistoryThrottledLogRPS // DecisionRetryCriticalAttempts is decision attempt threshold for logging and emiting metrics // KeyName: history.decisionRetryCriticalAttempts // Value type: Int // Default value: 10 // Allowed filters: N/A DecisionRetryCriticalAttempts // DecisionRetryMaxAttempts is the max limit for decision retry attempts. 0 indicates infinite number of attempts. // KeyName: history.decisionRetryMaxAttempts // Value type: Int // Default value: 1000 // Allowed filters: DomainName DecisionRetryMaxAttempts // NormalDecisionScheduleToStartMaxAttempts is the maximum decision attempt for creating a scheduleToStart timeout // timer for normal (non-sticky) decision // KeyName: history.normalDecisionScheduleToStartMaxAttempts // Value type: Int // Default value: 0 // Allowed filters: DomainName NormalDecisionScheduleToStartMaxAttempts // MaxBufferedQueryCount indicates the maximum number of queries which can be buffered at a given time for a single workflow // KeyName: history.MaxBufferedQueryCount // Value type: Int // Default value: 1 // Allowed filters: N/A MaxBufferedQueryCount // MutableStateChecksumGenProbability is the probability [0-100] that checksum will be generated for mutable state // KeyName: history.mutableStateChecksumGenProbability // Value type: Int // Default value: 0 // Allowed filters: DomainName MutableStateChecksumGenProbability // MutableStateChecksumVerifyProbability is the probability [0-100] that checksum will be verified for mutable state // KeyName: history.mutableStateChecksumVerifyProbability // Value type: Int // Default value: 0 // Allowed filters: DomainName MutableStateChecksumVerifyProbability // MaxActivityCountDispatchByDomain max # of activity tasks to dispatch to matching before creating transfer tasks. This is an performance optimization to skip activity scheduling efforts. // KeyName: history.activityDispatchForSyncMatchCountByDomain // Value type: Int // Default value: 0 // Allowed filters: DomainName MaxActivityCountDispatchByDomain // ReplicationTaskFetcherParallelism determines how many go routines we spin up for fetching tasks // KeyName: history.ReplicationTaskFetcherParallelism // Value type: Int // Default value: 1 // Allowed filters: N/A ReplicationTaskFetcherParallelism // ReplicationTaskProcessorErrorRetryMaxAttempts is the max retry attempts for applying replication tasks // KeyName: history.ReplicationTaskProcessorErrorRetryMaxAttempts // Value type: Int // Default value: 10 // Allowed filters: ShardID ReplicationTaskProcessorErrorRetryMaxAttempts // WorkflowIDExternalRPS is the rate limit per workflowID for external calls // KeyName: history.workflowIDExternalRPS // Value type: Int // Default value: UnlimitedRPS // Allowed filters: DomainName WorkflowIDExternalRPS // WorkflowIDInternalRPS is the rate limit per workflowID for internal calls // KeyName: history.workflowIDInternalRPS // Value type: Int // Default value: UnlimitedRPS // Allowed filters: DomainName WorkflowIDInternalRPS // WorkerPersistenceMaxQPS is the max qps worker host can query DB // KeyName: worker.persistenceMaxQPS // Value type: Int // Default value: 500 // Allowed filters: N/A WorkerPersistenceMaxQPS // WorkerPersistenceGlobalMaxQPS is the max qps worker cluster can query DB // KeyName: worker.persistenceGlobalMaxQPS // Value type: Int // Default value: 0 // Allowed filters: N/A WorkerPersistenceGlobalMaxQPS // WorkerIndexerConcurrency is the max concurrent messages to be processed at any given time // KeyName: worker.indexerConcurrency // Value type: Int // Default value: 1000 // Allowed filters: N/A WorkerIndexerConcurrency // WorkerESProcessorNumOfWorkers is num of workers for esProcessor // KeyName: worker.ESProcessorNumOfWorkers // Value type: Int // Default value: 1 // Allowed filters: N/A WorkerESProcessorNumOfWorkers // WorkerESProcessorBulkActions is max number of requests in bulk for esProcessor // KeyName: worker.ESProcessorBulkActions // Value type: Int // Default value: 1000 // Allowed filters: N/A WorkerESProcessorBulkActions // WorkerESProcessorBulkSize is max total size of bulk in bytes for esProcessor // KeyName: worker.ESProcessorBulkSize // Value type: Int // Default value: 2<<24 // 16MB // Allowed filters: N/A WorkerESProcessorBulkSize // WorkerArchiverConcurrency is controls the number of coroutines handling archival work per archival workflow // KeyName: worker.ArchiverConcurrency // Value type: Int // Default value: 50 // Allowed filters: N/A WorkerArchiverConcurrency // WorkerArchivalsPerIteration is controls the number of archivals handled in each iteration of archival workflow // KeyName: worker.ArchivalsPerIteration // Value type: Int // Default value: 1000 // Allowed filters: N/A WorkerArchivalsPerIteration // WorkerThrottledLogRPS is the rate limit on number of log messages emitted per second for throttled logger // KeyName: worker.throttledLogRPS // Value type: Int // Default value: 20 // Allowed filters: N/A WorkerThrottledLogRPS // ScannerPersistenceMaxQPS is the maximum rate of persistence calls from worker.Scanner // KeyName: worker.scannerPersistenceMaxQPS // Value type: Int // Default value: 5 // Allowed filters: N/A ScannerPersistenceMaxQPS // ScannerGetOrphanTasksPageSize is the maximum number of orphans to delete in one batch // KeyName: worker.scannerGetOrphanTasksPageSize // Value type: Int // Default value: 1000 // Allowed filters: N/A ScannerGetOrphanTasksPageSize // ScannerBatchSizeForTasklistHandler is for: 1. max number of tasks to query per call(get tasks for tasklist) in the scavenger handler. 2. The scavenger then uses the return to decide if a tasklist can be deleted. It's better to keep it a relatively high number to let it be more efficient. // KeyName: worker.scannerBatchSizeForTasklistHandler // Value type: Int // Default value: 1000 // Allowed filters: N/A ScannerBatchSizeForTasklistHandler // ScannerMaxTasksProcessedPerTasklistJob is the number of tasks to process for a tasklist in each workflow run // KeyName: worker.scannerMaxTasksProcessedPerTasklistJob // Value type: Int // Default value: 256 // Allowed filters: N/A ScannerMaxTasksProcessedPerTasklistJob // ConcreteExecutionsScannerConcurrency indicates the concurrency of concrete execution scanner // KeyName: worker.executionsScannerConcurrency // Value type: Int // Default value: 25 // Allowed filters: N/A ConcreteExecutionsScannerConcurrency // ConcreteExecutionsScannerBlobstoreFlushThreshold indicates the flush threshold of blobstore in concrete execution scanner // KeyName: worker.executionsScannerBlobstoreFlushThreshold // Value type: Int // Default value: 100 // Allowed filters: N/A ConcreteExecutionsScannerBlobstoreFlushThreshold // ConcreteExecutionsScannerActivityBatchSize indicates the batch size of scanner activities // KeyName: worker.executionsScannerActivityBatchSize // Value type: Int // Default value: 25 // Allowed filters: N/A ConcreteExecutionsScannerActivityBatchSize // ConcreteExecutionsScannerPersistencePageSize indicates the page size of execution persistence fetches in concrete execution scanner // KeyName: worker.executionsScannerPersistencePageSize // Value type: Int // Default value: 1000 // Allowed filters: N/A ConcreteExecutionsScannerPersistencePageSize // CurrentExecutionsScannerConcurrency indicates the concurrency of current executions scanner // KeyName: worker.currentExecutionsConcurrency // Value type: Int // Default value: 25 // Allowed filters: N/A CurrentExecutionsScannerConcurrency // CurrentExecutionsScannerBlobstoreFlushThreshold indicates the flush threshold of blobstore in current executions scanner // KeyName: worker.currentExecutionsBlobstoreFlushThreshold // Value type: Int // Default value: 100 // Allowed filters: N/A CurrentExecutionsScannerBlobstoreFlushThreshold // CurrentExecutionsScannerActivityBatchSize indicates the batch size of scanner activities // KeyName: worker.currentExecutionsActivityBatchSize // Value type: Int // Default value: 25 // Allowed filters: N/A CurrentExecutionsScannerActivityBatchSize // CurrentExecutionsScannerPersistencePageSize indicates the page size of execution persistence fetches in current executions scanner // KeyName: worker.currentExecutionsPersistencePageSize // Value type: INt // Default value: 1000 // Allowed filters: N/A CurrentExecutionsScannerPersistencePageSize // TimersScannerConcurrency is the concurrency of timers scanner // KeyName: worker.timersScannerConcurrency // Value type: Int // Default value: 5 // Allowed filters: N/A TimersScannerConcurrency // TimersScannerPersistencePageSize is the page size of timers persistence fetches in timers scanner // KeyName: worker.timersScannerPersistencePageSize // Value type: Int // Default value: 1000 // Allowed filters: N/A TimersScannerPersistencePageSize // TimersScannerBlobstoreFlushThreshold is threshold to flush blob store // KeyName: worker.timersScannerBlobstoreFlushThreshold // Value type: Int // Default value: 100 // Allowed filters: N/A TimersScannerBlobstoreFlushThreshold // TimersScannerActivityBatchSize is TimersScannerActivityBatchSize // KeyName: worker.timersScannerActivityBatchSize // Value type: Int // Default value: 25 // Allowed filters: N/A TimersScannerActivityBatchSize // TimersScannerPeriodStart is interval start for fetching scheduled timers // KeyName: worker.timersScannerPeriodStart // Value type: Int // Default value: 24 // Allowed filters: N/A TimersScannerPeriodStart // TimersScannerPeriodEnd is interval end for fetching scheduled timers // KeyName: worker.timersScannerPeriodEnd // Value type: Int // Default value: 3 // Allowed filters: N/A TimersScannerPeriodEnd // ESAnalyzerMaxNumDomains defines how many domains to check // KeyName: worker.ESAnalyzerMaxNumDomains // Value type: int // Default value: 500 ESAnalyzerMaxNumDomains // ESAnalyzerMaxNumWorkflowTypes defines how many workflow types to check per domain // KeyName: worker.ESAnalyzerMaxNumWorkflowTypes // Value type: int // Default value: 100 ESAnalyzerMaxNumWorkflowTypes // ESAnalyzerNumWorkflowsToRefresh controls how many workflows per workflow type should be refreshed per workflow type // KeyName: worker.ESAnalyzerNumWorkflowsToRefresh // Value type: Int // Default value: 100 ESAnalyzerNumWorkflowsToRefresh // ESAnalyzerMinNumWorkflowsForAvg controls how many workflows to have at least to rely on workflow run time avg per type // KeyName: worker.ESAnalyzerMinNumWorkflowsForAvg // Value type: Int // Default value: 100 ESAnalyzerMinNumWorkflowsForAvg // ShardManagerPersistenceMaxQPS is the max qps a shard manager host can query DB // KeyName: shardManager.persistenceMaxQPS // Value type: Int // Default value: 3000 // Allowed filters: N/A ShardManagerPersistenceMaxQPS // ShardManagerPersistenceGlobalMaxQPS is the max qps matching cluster can query DB // KeyName: shardManager.persistenceGlobalMaxQPS // Value type: Int // Default value: 0 // Allowed filters: N/A ShardManagerPersistenceGlobalMaxQPS // ShardManagerThrottledLogRPS is the rate limit on number of log messages emitted per second for throttled logger // KeyName: shardManager.throttledLogRPS // Value type: Int // Default value: 20 // Allowed filters: N/A ShardManagerThrottledLogRPS // Usage: VisibilityArchivalQueryMaxRangeInDays is the maximum number of days for a visibility archival query // KeyName: N/A // Default value: N/A // TODO: https://github.com/uber/cadence/issues/3861 VisibilityArchivalQueryMaxRangeInDays // Usage: VisibilityArchivalQueryMaxQPS is the timeout for a visibility archival query // KeyName: N/A // Default value: N/A // TODO: https://github.com/uber/cadence/issues/3861 VisibilityArchivalQueryMaxQPS // WorkflowDeletionJitterRange defines the duration in minutes for workflow close tasks jittering // KeyName: system.workflowDeletionJitterRange // Value type: Int // Default value: 1 (no jittering) WorkflowDeletionJitterRange // SampleLoggingRate defines the rate we want sampled logs to be logged at // KeyName: system.sampleLoggingRate // Value type: Int // Default value: 100 SampleLoggingRate // LargeShardHistorySizeMetricThreshold defines the threshold for what consititutes a large history storage size to alert on // KeyName: system.largeShardHistorySizeMetricThreshold // Value type: Int // Default value: 10485760 (10mb) LargeShardHistorySizeMetricThreshold // LargeShardHistoryEventMetricThreshold defines the threshold for what consititutes a large history event size to alert on // KeyName: system.largeShardHistoryEventMetricThreshold // Value type: Int // Default value: 50 * 1024 LargeShardHistoryEventMetricThreshold // IsolationGroupStateUpdateRetryAttempts // KeyName: system.isolationGroupStateUpdateRetryAttempts // Value type: int // Default value: 2 IsolationGroupStateUpdateRetryAttempts LargeShardHistoryBlobMetricThreshold // DeleteHistoryEventContextTimeout in seconds // KeyName: system.deleteHistoryEventContextTimeout // Value type: Int // Default value: 30 DeleteHistoryEventContextTimeout // LastIntKey must be the last one in this const group LastIntKey )
** * !!!Important!!! * For developer: Make sure to add/maintain the comment in the right format: usage, keyName, and default value * So that our go-docs can have the full [documentation](https://pkg.go.dev/github.com/uber/cadence@v0.19.1/common/service/dynamicconfig#Key). **
func (IntKey) DefaultInt ¶ added in v0.25.0
func (IntKey) DefaultValue ¶ added in v0.25.0
func (k IntKey) DefaultValue() interface{}
func (IntKey) Description ¶ added in v0.25.0
type IntPropertyFn ¶
type IntPropertyFn func(opts ...FilterOption) int
IntPropertyFn is a wrapper to get int property from dynamic config
func (IntPropertyFn) AsFloat64 ¶ added in v0.24.0
func (f IntPropertyFn) AsFloat64(opts ...FilterOption) func() float64
type IntPropertyFnWithDomainFilter ¶
IntPropertyFnWithDomainFilter is a wrapper to get int property from dynamic config with domain as filter
func (IntPropertyFnWithDomainFilter) AsFloat64 ¶ added in v0.24.0
func (f IntPropertyFnWithDomainFilter) AsFloat64(domain string) func() float64
type IntPropertyFnWithShardIDFilter ¶
IntPropertyFnWithShardIDFilter is a wrapper to get int property from dynamic config with shardID as filter
type IntPropertyFnWithTaskListInfoFilters ¶
IntPropertyFnWithTaskListInfoFilters is a wrapper to get int property from dynamic config with three filters: domain, taskList, taskType
type IntPropertyFnWithWorkflowTypeFilter ¶ added in v0.24.0
IntPropertyFnWithWorkflowTypeFilter is a wrapper to get int property from dynamic config with domain as filter
type Key ¶
type Key interface { String() string Description() string DefaultValue() interface{} // Filters is used to identify what filters a DynamicConfig key may have. // For example, CLI tool uses this to figure out all domain specific configurations for migration validation. Filters() []Filter }
func GetKeyFromKeyName ¶ added in v0.25.0
func ListAllProductionKeys ¶ added in v0.25.0
func ListAllProductionKeys() []Key
ListAllProductionKeys returns all key used in production
type ListKey ¶ added in v1.0.0
type ListKey int
const ( UnknownListKey ListKey = iota TestGetListPropertyKey // AllIsolationGroups is the list of all possible isolation groups in a service // KeyName: system.allIsolationGroups // Value type: []string // Default value: N/A // Allowed filters: N/A AllIsolationGroups // HeaderForwardingRules defines which headers are forwarded from inbound calls to outbound. // This value is only loaded at startup. // // Regexes and header names are used as-is, you are strongly encouraged to use `(?i)` to make your regex case-insensitive. // // KeyName: admin.HeaderForwardingRules // Value type: []rpc.HeaderRule or an []interface{} containing `map[string]interface{}{"Add":bool,"Match":string}` values. // Default value: forward all headers. (this is a problematic value, and it will be changing as we reduce to a list of known values) HeaderForwardingRules LastListKey )
const DefaultIsolationGroupConfigStoreManagerGlobalMapping ListKey = -1 // This is a hack to put it in a different list due to it being a different config type
DefaultIsolationGroupConfigStoreManagerGlobalMapping is the dynamic config value for isolation groups Note: This is not typically used for normal dynamic config (type 0), but instead it's used only for IsolationGroup config (type 1). KeyName: system.defaultIsolationGroupConfigStoreManagerGlobalMapping
func (ListKey) DefaultList ¶ added in v1.0.0
func (k ListKey) DefaultList() []interface{}
func (ListKey) DefaultValue ¶ added in v1.0.0
func (k ListKey) DefaultValue() interface{}
func (ListKey) Description ¶ added in v1.0.0
type ListPropertyFn ¶ added in v1.0.0
type ListPropertyFn func(opts ...FilterOption) []interface{}
ListPropertyFn is a wrapper to get a list property from dynamic config
type MapKey ¶ added in v0.25.0
type MapKey int
const ( UnknownMapKey MapKey = iota // key for tests TestGetMapPropertyKey // RequiredDomainDataKeys is the key for the list of data keys required in domain registration // KeyName: system.requiredDomainDataKeys // Value type: Map // Default value: nil // Allowed filters: N/A RequiredDomainDataKeys // ValidSearchAttributes is legal indexed keys that can be used in list APIs. When overriding, ensure to include the existing default attributes of the current release // KeyName: frontend.validSearchAttributes // Value type: Map // Default value: the default attributes of this release version, see definition.GetDefaultIndexedKeys() // Allowed filters: N/A ValidSearchAttributes // TaskSchedulerRoundRobinWeights is the priority weight for weighted round robin task scheduler // KeyName: history.taskSchedulerRoundRobinWeight // Value type: Map // Default value: please see common.ConvertIntMapToDynamicConfigMapProperty(DefaultTaskPriorityWeight) in code base // Allowed filters: N/A TaskSchedulerRoundRobinWeights // QueueProcessorPendingTaskSplitThreshold is the threshold for the number of pending tasks per domain // KeyName: history.queueProcessorPendingTaskSplitThreshold // Value type: Map // Default value: see common.ConvertIntMapToDynamicConfigMapProperty(DefaultPendingTaskSplitThreshold) in code base // Allowed filters: N/A QueueProcessorPendingTaskSplitThreshold // QueueProcessorStuckTaskSplitThreshold is the threshold for the number of attempts of a task // KeyName: history.queueProcessorStuckTaskSplitThreshold // Value type: Map // Default value: see common.ConvertIntMapToDynamicConfigMapProperty(DefaultStuckTaskSplitThreshold) in code base // Allowed filters: N/A QueueProcessorStuckTaskSplitThreshold // LastMapKey must be the last one in this const group LastMapKey )
func (MapKey) DefaultMap ¶ added in v0.25.0
func (MapKey) DefaultValue ¶ added in v0.25.0
func (k MapKey) DefaultValue() interface{}
func (MapKey) Description ¶ added in v0.25.0
type MapPropertyFn ¶
type MapPropertyFn func(opts ...FilterOption) map[string]interface{}
MapPropertyFn is a wrapper to get map property from dynamic config
type MockClient ¶
type MockClient struct {
// contains filtered or unexported fields
}
MockClient is a mock of Client interface.
func NewMockClient ¶
func NewMockClient(ctrl *gomock.Controller) *MockClient
NewMockClient creates a new mock instance.
func (*MockClient) EXPECT ¶
func (m *MockClient) EXPECT() *MockClientMockRecorder
EXPECT returns an object that allows the caller to indicate expected use.
func (*MockClient) GetBoolValue ¶
func (m *MockClient) GetBoolValue(name BoolKey, filters map[Filter]interface{}) (bool, error)
GetBoolValue mocks base method.
func (*MockClient) GetDurationValue ¶
func (m *MockClient) GetDurationValue(name DurationKey, filters map[Filter]interface{}) (time.Duration, error)
GetDurationValue mocks base method.
func (*MockClient) GetFloatValue ¶
func (m *MockClient) GetFloatValue(name FloatKey, filters map[Filter]interface{}) (float64, error)
GetFloatValue mocks base method.
func (*MockClient) GetIntValue ¶
func (m *MockClient) GetIntValue(name IntKey, filters map[Filter]interface{}) (int, error)
GetIntValue mocks base method.
func (*MockClient) GetListValue ¶ added in v1.0.0
func (m *MockClient) GetListValue(name ListKey, filters map[Filter]interface{}) ([]interface{}, error)
GetListValue mocks base method.
func (*MockClient) GetMapValue ¶
func (m *MockClient) GetMapValue(name MapKey, filters map[Filter]interface{}) (map[string]interface{}, error)
GetMapValue mocks base method.
func (*MockClient) GetStringValue ¶
func (m *MockClient) GetStringValue(name StringKey, filters map[Filter]interface{}) (string, error)
GetStringValue mocks base method.
func (*MockClient) GetValue ¶
func (m *MockClient) GetValue(name Key) (interface{}, error)
GetValue mocks base method.
func (*MockClient) GetValueWithFilters ¶
func (m *MockClient) GetValueWithFilters(name Key, filters map[Filter]interface{}) (interface{}, error)
GetValueWithFilters mocks base method.
func (*MockClient) ListValue ¶ added in v0.23.1
func (m *MockClient) ListValue(name Key) ([]*types.DynamicConfigEntry, error)
ListValue mocks base method.
func (*MockClient) RestoreValue ¶ added in v0.23.1
func (m *MockClient) RestoreValue(name Key, filters map[Filter]interface{}) error
RestoreValue mocks base method.
func (*MockClient) UpdateValue ¶
func (m *MockClient) UpdateValue(name Key, value interface{}) error
UpdateValue mocks base method.
type MockClientMockRecorder ¶
type MockClientMockRecorder struct {
// contains filtered or unexported fields
}
MockClientMockRecorder is the mock recorder for MockClient.
func (*MockClientMockRecorder) GetBoolValue ¶
func (mr *MockClientMockRecorder) GetBoolValue(name, filters interface{}) *gomock.Call
GetBoolValue indicates an expected call of GetBoolValue.
func (*MockClientMockRecorder) GetDurationValue ¶
func (mr *MockClientMockRecorder) GetDurationValue(name, filters interface{}) *gomock.Call
GetDurationValue indicates an expected call of GetDurationValue.
func (*MockClientMockRecorder) GetFloatValue ¶
func (mr *MockClientMockRecorder) GetFloatValue(name, filters interface{}) *gomock.Call
GetFloatValue indicates an expected call of GetFloatValue.
func (*MockClientMockRecorder) GetIntValue ¶
func (mr *MockClientMockRecorder) GetIntValue(name, filters interface{}) *gomock.Call
GetIntValue indicates an expected call of GetIntValue.
func (*MockClientMockRecorder) GetListValue ¶ added in v1.0.0
func (mr *MockClientMockRecorder) GetListValue(name, filters interface{}) *gomock.Call
GetListValue indicates an expected call of GetListValue.
func (*MockClientMockRecorder) GetMapValue ¶
func (mr *MockClientMockRecorder) GetMapValue(name, filters interface{}) *gomock.Call
GetMapValue indicates an expected call of GetMapValue.
func (*MockClientMockRecorder) GetStringValue ¶
func (mr *MockClientMockRecorder) GetStringValue(name, filters interface{}) *gomock.Call
GetStringValue indicates an expected call of GetStringValue.
func (*MockClientMockRecorder) GetValue ¶
func (mr *MockClientMockRecorder) GetValue(name interface{}) *gomock.Call
GetValue indicates an expected call of GetValue.
func (*MockClientMockRecorder) GetValueWithFilters ¶
func (mr *MockClientMockRecorder) GetValueWithFilters(name, filters interface{}) *gomock.Call
GetValueWithFilters indicates an expected call of GetValueWithFilters.
func (*MockClientMockRecorder) ListValue ¶ added in v0.23.1
func (mr *MockClientMockRecorder) ListValue(name interface{}) *gomock.Call
ListValue indicates an expected call of ListValue.
func (*MockClientMockRecorder) RestoreValue ¶ added in v0.23.1
func (mr *MockClientMockRecorder) RestoreValue(name, filters interface{}) *gomock.Call
RestoreValue indicates an expected call of RestoreValue.
func (*MockClientMockRecorder) UpdateValue ¶
func (mr *MockClientMockRecorder) UpdateValue(name, value interface{}) *gomock.Call
UpdateValue indicates an expected call of UpdateValue.
type PropertyFn ¶
type PropertyFn func() interface{}
PropertyFn is a wrapper to get property from dynamic config
type StringKey ¶ added in v0.25.0
type StringKey int
const ( UnknownStringKey StringKey = iota // key for tests TestGetStringPropertyKey // AdvancedVisibilityWritingMode is key for how to write to advanced visibility. The most useful option is "dual", which can be used for seamless migration from db visibility to advanced visibility, usually using with EnableReadVisibilityFromES // KeyName: system.advancedVisibilityWritingMode // Value type: String enum: "on"(means writing to advancedVisibility only, "off" (means writing to db visibility only), or "dual" (means writing to both) // Default value: "on" // Allowed filters: N/A AdvancedVisibilityWritingMode // AdvancedVisibilityMigrationWritingMode is key for how to write to advanced visibility during migration. // KeyName: system.AdvancedVisibilityMigrationWritingMode // Value type: String enum: "dual"(means writing to both source and destination advanced visibility, "source" (means writing to source visibility only), "destination" (means writing to destination visibility only) or "off" (means writing to db visibility only) // Default value: "dual" // Allowed filters: N/A AdvancedVisibilityMigrationWritingMode // HistoryArchivalStatus is key for the status of history archival to override the value from static config. // KeyName: system.historyArchivalStatus // Value type: string enum: "enabled" or "disabled" // Default value: "enabled" // Allowed filters: N/A HistoryArchivalStatus // VisibilityArchivalStatus is key for the status of visibility archival to override the value from static config. // KeyName: system.visibilityArchivalStatus // Value type: string enum: "enabled" or "disabled" // Default value: "enabled" // Allowed filters: N/A VisibilityArchivalStatus // DefaultEventEncoding is the encoding type for history events // KeyName: history.defaultEventEncoding // Value type: String // Default value: string(common.EncodingTypeThriftRW) // Allowed filters: DomainName DefaultEventEncoding // AdminOperationToken is the token to pass admin checking // KeyName: history.adminOperationToken // Value type: String // Default value: common.DefaultAdminOperationToken // Allowed filters: N/A AdminOperationToken // ESAnalyzerLimitToTypes controls if we want to limit ESAnalyzer only to some workflow types // KeyName: worker.ESAnalyzerLimitToTypes // Value type: String // Default value: "" => means no limitation ESAnalyzerLimitToTypes // ESAnalyzerLimitToDomains controls if we want to limit ESAnalyzer only to some domains // KeyName: worker.ESAnalyzerLimitToDomains // Value type: String // Default value: "" => means no limitation ESAnalyzerLimitToDomains // ESAnalyzerWorkflowDurationWarnThresholds defines the warning execution thresholds for workflow types // KeyName: worker.ESAnalyzerWorkflowDurationWarnThresholds // Value type: string [{"DomainName":"<domain>", "WorkflowType":"<workflowType>", "Threshold":"<duration>", "Refresh":<shouldRefresh>, "MaxNumWorkflows":<maxNumber>}] // Default value: "" ESAnalyzerWorkflowDurationWarnThresholds // ESAnalyzerWorkflowVersionMetricDomains defines the domains we want to emit wf version metrics on // KeyName: worker.ESAnalyzerWorkflowVersionMetricDomains // Value type: string ["test-domain","test-domain2"] // Default value: "" ESAnalyzerWorkflowVersionMetricDomains // ESAnalyzerWorkflowTypeMetricDomains defines the domains we want to emit wf type metrics on // KeyName: worker.ESAnalyzerWorkflowTypeMetricDomains // Value type: string ["test-domain","test-domain2"] // Default value: "" ESAnalyzerWorkflowTypeMetricDomains // FrontendGlobalRatelimiterMode controls what keys use global vs fallback behavior, // and whether shadowing is enabled. This is only available for frontend usage for now. // // - "disabled" stops usage-tracking and all Update requests, in an attempt to be as close to "do not use at all" as possible. // - "local" uses the new limiters with call tracking and metrics, but forces local-only behavior and does not submit usage data to aggregators. // - "global" uses the new global-load-balanced logic (though it may decide to use a local-fallback internally, and this is not prevented) // - "x-shadow-y" means "use x, and shadow all calls to y but ignore the result". // this calls both, tracks and emits both metrics, and can be used to "warm" either limiter's in-memory state before switching. // // These values can be seen as constants of github.com/uber/cadence/common/quotas/global/collection.keyMode // // KeyName: frontend.globalRatelimiterMode // Value type: string enum: "disabled", "local", "global", "local-shadow-global", or "global-shadow-local" // Default value: "disabled" // Allowed filters: RatelimitKey (on global key, e.g. prefixed by collection name) FrontendGlobalRatelimiterMode TasklistLoadBalancerStrategy // LastStringKey must be the last one in this const group LastStringKey )
func (StringKey) DefaultString ¶ added in v0.25.0
func (StringKey) DefaultValue ¶ added in v0.25.0
func (k StringKey) DefaultValue() interface{}
func (StringKey) Description ¶ added in v0.25.0
type StringPropertyFn ¶
type StringPropertyFn func(opts ...FilterOption) string
StringPropertyFn is a wrapper to get string property from dynamic config
type StringPropertyFnWithDomainFilter ¶
StringPropertyFnWithDomainFilter is a wrapper to get string property from dynamic config
type StringPropertyFnWithTaskListInfoFilters ¶ added in v1.2.14
type StringPropertyFnWithTaskListInfoFilters func(domain string, taskList string, taskType int) string
StringPropertyFnWithTaskListInfoFilters is a wrapper to get string property from dynamic config with domainID as filter
type StringPropertyWithRatelimitKeyFilter ¶ added in v1.2.12
StringPropertyWithRatelimitKeyFilter is a wrapper to get strings (currently global ratelimiter modes) per global ratelimit key
Source Files ¶
Directories ¶
Path | Synopsis |
---|---|
Package configstore is a generated GoMock package.
|
Package configstore is a generated GoMock package. |