worker

package
v1.5.2 Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Jan 22, 2021 License: MIT Imports: 18 Imported by: 0

README

Temporal Worker

Temporal Worker is a role for Temporal service used for hosting any components responsible for performing background processing on the Temporal cluster.

Replicator

Replicator is a background worker responsible for consuming replication tasks generated by remote Temporal clusters and pass it down to processor so they can be applied to local Temporal cluster.

It uses Kafka as the replication tasks buffer and relies on [kafka-client library] (https://github.com/temporalio/kafka-client/) for consuming messages from Kafka.

Quickstart for localhost development

  1. Setup Kafka by following instructions: Kafka Quickstart
  2. Create Kafka topic for active and standby clusters if needed. By default the development Kafka should create topics in- flight (with 1 partition). If not, then use the follow command to create topics:
bin/kafka-topics.sh --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic active

and

bin/kafka-topics.sh --create --zookeeper localhost:2181 --replication-factor 1 --partitions 1 --topic standby
  1. Start Temporal development server for active zone:
./temporal-server --zone active start
  1. Start Temporal development server for standby(passive) zone:
./temporal-server --zone standby start
  1. Create global namespaces
tctl --ns sample namespace register --gd true --ac active --cl active standby
  1. Failover between zones:

Failover to standby:

tctl --ns sample namespace update --ac standby

Failback to active:

tctl --ns sample namespace update --ac active

Create replication task using CLI

Kafka CLI can be used to generate a replication task for testing purpose:

bin/kafka-console-producer.sh --broker-list localhost:9092 --topic standby

Replication task message:

{taskType: 0}

Archiver

Archiver is used to handle archival of workflow execution histories. It does this by hosting a temporal client worker and running an archival system workflow. The archival client gets used to initiate archival through signal sending. The archiver shards work across several workflows.

Documentation

Index

Constants

This section is empty.

Variables

This section is empty.

Functions

func NewService

func NewService(
	params *resource.BootstrapParams,
) (resource.Resource, error)

NewService builds a new worker service

Types

type Config

type Config struct {
	ReplicationCfg                *replicator.Config
	ArchiverConfig                *archiver.Config
	IndexerCfg                    *indexer.Config
	ScannerCfg                    *scanner.Config
	BatcherCfg                    *batcher.Config
	ThrottledLogRPS               dynamicconfig.IntPropertyFn
	PersistenceGlobalMaxQPS       dynamicconfig.IntPropertyFn
	EnableBatcher                 dynamicconfig.BoolPropertyFn
	VisibilityQueue               dynamicconfig.StringPropertyFn
	EnableParentClosePolicyWorker dynamicconfig.BoolPropertyFn
}

Config contains all the service config for worker

func NewConfig

func NewConfig(params *resource.BootstrapParams) *Config

NewConfig builds the new Config for worker service

type Service

type Service struct {
	resource.Resource
	// contains filtered or unexported fields
}

Service represents the temporal-worker service. This service hosts all background processing needed for temporal cluster: 1. Replicator: Handles applying replication tasks generated by remote clusters. 2. Indexer: Handles uploading of visibility records to elastic search. 3. Archiver: Handles archival of workflow histories.

func (*Service) Start

func (s *Service) Start()

Start is called to start the service

func (*Service) Stop

func (s *Service) Stop()

Stop is called to stop the service

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL