Documentation ¶
Overview ¶
Package leaderelection implements leader election of a set of endpoints. It uses an annotation in the endpoints object to store the record of the election state. This implementation does not guarantee that only one client is acting as a leader (a.k.a. fencing).
A client only acts on timestamps captured locally to infer the state of the leader election. The client does not consider timestamps in the leader election record to be accurate because these timestamps may not have been produced by a local clock. The implemention does not depend on their accuracy and only uses their change to indicate that another client has renewed the leader lease. Thus the implementation is tolerant to arbitrary clock skew, but is not tolerant to arbitrary clock skew rate.
However the level of tolerance to skew rate can be configured by setting RenewDeadline and LeaseDuration appropriately. The tolerance expressed as a maximum tolerated ratio of time passed on the fastest node to time passed on the slowest node can be approximately achieved with a configuration that sets the same ratio of LeaseDuration to RenewDeadline. For example if a user wanted to tolerate some nodes progressing forward in time twice as fast as other nodes, the user could set LeaseDuration to 60 seconds and RenewDeadline to 30 seconds.
While not required, some method of clock synchronization between nodes in the cluster is highly recommended. It's important to keep in mind when configuring this client that the tolerance to skew rate varies inversely to master availability.
Larger clusters often have a more lenient SLA for API latency. This should be taken into account when configuring the client. The rate of leader transitions should be monitored and RetryPeriod and LeaseDuration should be increased until the rate is stable and acceptably low. It's important to keep in mind when configuring this client that the tolerance to API latency varies inversely to master availability.
DISCLAIMER: this is an alpha API. This library will likely change significantly or even be removed entirely in subsequent releases. Depend on this API at your own risk.
Index ¶
- Constants
- func NewCandidate(clientset kubernetes.Interface, candidateNamespace string, ...) (*LeaseCandidate, CacheSyncWaiter, error)
- func RunOrDie(ctx context.Context, lec LeaderElectionConfig)
- func SetProvider(metricsProvider MetricsProvider)
- type CacheSyncWaiter
- type HealthzAdaptor
- type LeaderCallbacks
- type LeaderElectionConfig
- type LeaderElector
- type LeaderMetric
- type LeaseCandidate
- type MetricsProvider
Constants ¶
const (
JitterFactor = 1.2
)
Variables ¶
This section is empty.
Functions ¶
func NewCandidate ¶ added in v0.31.0
func NewCandidate(clientset kubernetes.Interface, candidateNamespace string, candidateName string, targetLease string, binaryVersion, emulationVersion string, preferredStrategies []v1.CoordinatedLeaseStrategy, ) (*LeaseCandidate, CacheSyncWaiter, error)
NewCandidate creates new LeaseCandidate controller that creates a LeaseCandidate object if it does not exist and watches changes to the corresponding object and renews if PingTime is set. WARNING: This is an ALPHA feature. Ensure that the CoordinatedLeaderElection feature gate is on.
func RunOrDie ¶
func RunOrDie(ctx context.Context, lec LeaderElectionConfig)
RunOrDie starts a client with the provided config or panics if the config fails to validate. RunOrDie blocks until leader election loop is stopped by ctx or it has stopped holding the leader lease
func SetProvider ¶
func SetProvider(metricsProvider MetricsProvider)
SetProvider sets the metrics provider for all subsequently created work queues. Only the first call has an effect.
Types ¶
type CacheSyncWaiter ¶ added in v0.31.0
type HealthzAdaptor ¶
type HealthzAdaptor struct {
// contains filtered or unexported fields
}
HealthzAdaptor associates the /healthz endpoint with the LeaderElection object. It helps deal with the /healthz endpoint being set up prior to the LeaderElection. This contains the code needed to act as an adaptor between the leader election code the health check code. It allows us to provide health status about the leader election. Most specifically about if the leader has failed to renew without exiting the process. In that case we should report not healthy and rely on the kubelet to take down the process.
func NewLeaderHealthzAdaptor ¶
func NewLeaderHealthzAdaptor(timeout time.Duration) *HealthzAdaptor
NewLeaderHealthzAdaptor creates a basic healthz adaptor to monitor a leader election. timeout determines the time beyond the lease expiry to be allowed for timeout. checks within the timeout period after the lease expires will still return healthy.
func (*HealthzAdaptor) Check ¶
func (l *HealthzAdaptor) Check(req *http.Request) error
Check is called by the healthz endpoint handler. It fails (returns an error) if we own the lease but had not been able to renew it.
func (*HealthzAdaptor) Name ¶
func (l *HealthzAdaptor) Name() string
Name returns the name of the health check we are implementing.
func (*HealthzAdaptor) SetLeaderElection ¶
func (l *HealthzAdaptor) SetLeaderElection(le *LeaderElector)
SetLeaderElection ties a leader election object to a HealthzAdaptor
type LeaderCallbacks ¶
type LeaderCallbacks struct { // OnStartedLeading is called when a LeaderElector client starts leading OnStartedLeading func(context.Context) // OnStoppedLeading is called when a LeaderElector client stops leading. // This callback is always called when the LeaderElector exits, even if it did not start leading. // Users should not assume that OnStoppedLeading is only called after OnStartedLeading. // see: https://github.com/kubernetes/kubernetes/pull/127675#discussion_r1780059887 OnStoppedLeading func() // OnNewLeader is called when the client observes a leader that is // not the previously observed leader. This includes the first observed // leader when the client starts. OnNewLeader func(identity string) }
LeaderCallbacks are callbacks that are triggered during certain lifecycle events of the LeaderElector. These are invoked asynchronously.
possible future callbacks:
- OnChallenge()
type LeaderElectionConfig ¶
type LeaderElectionConfig struct { // Lock is the resource that will be used for locking Lock rl.Interface // LeaseDuration is the duration that non-leader candidates will // wait to force acquire leadership. This is measured against time of // last observed ack. // // A client needs to wait a full LeaseDuration without observing a change to // the record before it can attempt to take over. When all clients are // shutdown and a new set of clients are started with different names against // the same leader record, they must wait the full LeaseDuration before // attempting to acquire the lease. Thus LeaseDuration should be as short as // possible (within your tolerance for clock skew rate) to avoid a possible // long waits in the scenario. // // Core clients default this value to 15 seconds. LeaseDuration time.Duration // RenewDeadline is the duration that the acting master will retry // refreshing leadership before giving up. // // Core clients default this value to 10 seconds. RenewDeadline time.Duration // RetryPeriod is the duration the LeaderElector clients should wait // between tries of actions. // // Core clients default this value to 2 seconds. RetryPeriod time.Duration // Callbacks are callbacks that are triggered during certain lifecycle // events of the LeaderElector Callbacks LeaderCallbacks // WatchDog is the associated health checker // WatchDog may be null if it's not needed/configured. WatchDog *HealthzAdaptor // ReleaseOnCancel should be set true if the lock should be released // when the run context is cancelled. If you set this to true, you must // ensure all code guarded by this lease has successfully completed // prior to cancelling the context, or you may have two processes // simultaneously acting on the critical path. ReleaseOnCancel bool // Name is the name of the resource lock for debugging Name string // Coordinated will use the Coordinated Leader Election feature // WARNING: Coordinated leader election is ALPHA. Coordinated bool }
type LeaderElector ¶
type LeaderElector struct {
// contains filtered or unexported fields
}
LeaderElector is a leader election client.
func NewLeaderElector ¶
func NewLeaderElector(lec LeaderElectionConfig) (*LeaderElector, error)
NewLeaderElector creates a LeaderElector from a LeaderElectionConfig
func (*LeaderElector) Check ¶
func (le *LeaderElector) Check(maxTolerableExpiredLease time.Duration) error
Check will determine if the current lease is expired by more than timeout.
func (*LeaderElector) GetLeader ¶
func (le *LeaderElector) GetLeader() string
GetLeader returns the identity of the last observed leader or returns the empty string if no leader has yet been observed. This function is for informational purposes. (e.g. monitoring, logs, etc.)
func (*LeaderElector) IsLeader ¶
func (le *LeaderElector) IsLeader() bool
IsLeader returns true if the last observed leader was this client else returns false.
func (*LeaderElector) Run ¶
func (le *LeaderElector) Run(ctx context.Context)
Run starts the leader election loop. Run will not return before leader election loop is stopped by ctx or it has stopped holding the leader lease
type LeaderMetric ¶ added in v0.30.0
LeaderMetric instruments metrics used in leader election.
type LeaseCandidate ¶ added in v0.31.0
type LeaseCandidate struct {
// contains filtered or unexported fields
}
func (*LeaseCandidate) Run ¶ added in v0.31.0
func (c *LeaseCandidate) Run(ctx context.Context)
type MetricsProvider ¶
type MetricsProvider interface {
NewLeaderMetric() LeaderMetric
}
MetricsProvider generates various metrics used by the leader election.