store

package
v0.17.0 Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Nov 18, 2020 License: Apache-2.0 Imports: 66 Imported by: 18

Documentation

Index

Constants

View Source
const (
	// MaxSamplesPerChunk is approximately the max number of samples that we may have in any given chunk. This is needed
	// for precalculating the number of samples that we may have to retrieve and decode for any given query
	// without downloading them. Please take a look at https://github.com/prometheus/tsdb/pull/397 to know
	// where this number comes from. Long story short: TSDB is made in such a way, and it is made in such a way
	// because you barely get any improvements in compression when the number of samples is beyond this.
	// Take a look at Figure 6 in this whitepaper http://www.vldb.org/pvldb/vol8/p1816-teller.pdf.
	MaxSamplesPerChunk = 120

	// CompatibilityTypeLabelName is an artificial label that Store Gateway can optionally advertise. This is required for compatibility
	// with pre v0.8.0 Querier. Previous Queriers was strict about duplicated external labels of all StoreAPIs that had any labels.
	// Now with newer Store Gateway advertising all the external labels it has access to, there was simple case where
	// Querier was blocking Store Gateway as duplicate with sidecar.
	//
	// Newer Queriers are not strict, no duplicated external labels check is there anymore.
	// Additionally newer Queriers removes/ignore this exact labels from UI and querying.
	//
	// This label name is intentionally against Prometheus label style.
	// TODO(bwplotka): Remove it at some point.
	CompatibilityTypeLabelName = "@thanos_compatibility_store_type"

	// DefaultPostingOffsetInMemorySampling represents default value for --store.index-header-posting-offsets-in-mem-sampling.
	// 32 value is chosen as it's a good balance for common setups. Sampling that is not too large (too many CPU cycles) and
	// not too small (too much memory).
	DefaultPostingOffsetInMemorySampling = 32
)
View Source
const StoreMatcherKey = ctxKey(0)

StoreMatcherKey is the context key for the store's allow list.

Variables

This section is empty.

Functions

func NewNoCopyScanner added in v0.13.0

func NewNoCopyScanner(b []byte, splitFunc bufio.SplitFunc) *noCopyScanner

NewNoCopyScanner returns bufio.Scanner-like scanner that is meant to be used on already allocated byte slice (or mmapped) one. Returned tokens are shared.

func RegisterStoreServer added in v0.16.0

func RegisterStoreServer(storeSrv storepb.StoreServer) func(*grpc.Server)

func RegisterWritableStoreServer added in v0.16.0

func RegisterWritableStoreServer(storeSrv storepb.WriteableStoreServer) func(*grpc.Server)

func ScanGRPCCurlProtoStreamMessages added in v0.13.0

func ScanGRPCCurlProtoStreamMessages(data []byte, atEOF bool) (advance int, token []byte, err error)

ScanGRPCCurlProtoStreamMessages allows to tokenize each streamed gRPC message from grpcurl tool.

Types

type BucketStore

type BucketStore struct {
	// contains filtered or unexported fields
}

BucketStore implements the store API backed by a bucket. It loads all index files to local disk.

NOTE: Bucket store reencodes postings using diff+varint+snappy when storing to cache. This makes them smaller, but takes extra CPU and memory. When used with in-memory cache, memory usage should decrease overall, thanks to postings being smaller.

func NewBucketStore

func NewBucketStore(
	logger log.Logger,
	reg prometheus.Registerer,
	bkt objstore.InstrumentedBucketReader,
	fetcher block.MetadataFetcher,
	dir string,
	indexCache storecache.IndexCache,
	queryGate gate.Gate,
	maxChunkPoolBytes uint64,
	chunksLimiterFactory ChunksLimiterFactory,
	debugLogging bool,
	blockSyncConcurrency int,
	filterConfig *FilterConfig,
	enableCompatibilityLabel bool,
	postingOffsetsInMemSampling int,
	enableSeriesResponseHints bool,
	lazyIndexReaderEnabled bool,
	lazyIndexReaderIdleTimeout time.Duration,
) (*BucketStore, error)

NewBucketStore creates a new bucket backed store that implements the store API against an object store bucket. It is optimized to work against high latency backends.

func (*BucketStore) Close

func (s *BucketStore) Close() (err error)

Close the store.

func (*BucketStore) Info

Info implements the storepb.StoreServer interface.

func (*BucketStore) InitialSync

func (s *BucketStore) InitialSync(ctx context.Context) error

InitialSync perform blocking sync with extra step at the end to delete locally saved blocks that are no longer present in the bucket. The mismatch of these can only happen between restarts, so we can do that only once per startup.

func (*BucketStore) LabelNames

LabelNames implements the storepb.StoreServer interface.

func (*BucketStore) LabelValues

LabelValues implements the storepb.StoreServer interface.

func (*BucketStore) Series

func (s *BucketStore) Series(req *storepb.SeriesRequest, srv storepb.Store_SeriesServer) (err error)

Series implements the storepb.StoreServer interface.

func (*BucketStore) SyncBlocks

func (s *BucketStore) SyncBlocks(ctx context.Context) error

SyncBlocks synchronizes the stores state with the Bucket bucket. It will reuse disk space as persistent cache based on s.dir param.

func (*BucketStore) TimeRange

func (s *BucketStore) TimeRange() (mint, maxt int64)

TimeRange returns the minimum and maximum timestamp of data available in the store.

type ChunksLimiter added in v0.15.0

type ChunksLimiter interface {
	// Reserve num chunks out of the total number of chunks enforced by the limiter.
	// Returns an error if the limit has been exceeded. This function must be
	// goroutine safe.
	Reserve(num uint64) error
}

type ChunksLimiterFactory added in v0.15.0

type ChunksLimiterFactory func(failedCounter prometheus.Counter) ChunksLimiter

ChunksLimiterFactory is used to create a new ChunksLimiter. The factory is useful for projects depending on Thanos (eg. Cortex) which have dynamic limits.

func NewChunksLimiterFactory added in v0.15.0

func NewChunksLimiterFactory(limit uint64) ChunksLimiterFactory

NewChunksLimiterFactory makes a new ChunksLimiterFactory with a static limit.

type Client

type Client interface {
	// Client to access the store.
	storepb.StoreClient

	// LabelSets that each apply to some data exposed by the backing store.
	LabelSets() []labels.Labels

	// Minimum and maximum time range of data in the store.
	TimeRange() (mint int64, maxt int64)

	String() string
	// Addr returns address of a Client.
	Addr() string
}

Client holds meta information about a store.

type CloseDelegator added in v0.15.0

type CloseDelegator interface {
	Delegate(io.Closer)
}

CloseDelegator allows to delegate close (releasing resources used by request to the server). This is useful when we invoke StoreAPI within another StoreAPI and results are ephemeral until copied.

type FilterConfig added in v0.7.0

type FilterConfig struct {
	MinTime, MaxTime model.TimeOrDurationValue
}

FilterConfig is a configuration, which Store uses for filtering metrics based on time.

type Limiter added in v0.7.0

type Limiter struct {
	// contains filtered or unexported fields
}

Limiter is a simple mechanism for checking if something has passed a certain threshold.

func NewLimiter added in v0.7.0

func NewLimiter(limit uint64, ctr prometheus.Counter) *Limiter

NewLimiter returns a new limiter with a specified limit. 0 disables the limit.

func (*Limiter) Reserve added in v0.15.0

func (l *Limiter) Reserve(num uint64) error

Reserve implements ChunksLimiter.

type LocalStore added in v0.13.0

type LocalStore struct {
	// contains filtered or unexported fields
}

LocalStore implements the store API against single file with stream of proto-based SeriesResponses in JSON format. Inefficient implementation for quick StoreAPI view. Chunk order is exactly the same as in a given file.

func NewLocalStoreFromJSONMmappableFile added in v0.13.0

func NewLocalStoreFromJSONMmappableFile(
	logger log.Logger,
	component component.StoreAPI,
	extLabels labels.Labels,
	path string,
	split bufio.SplitFunc,
) (*LocalStore, error)

TODO(bwplotka): Add remote read so Prometheus users can use this. Potentially after streaming will be added https://github.com/prometheus/prometheus/issues/5926. TODO(bwplotka): Consider non mmaped version of this, as well different versions.

func (*LocalStore) Close added in v0.13.0

func (s *LocalStore) Close() (err error)

func (*LocalStore) Info added in v0.13.0

Info returns store information about the Prometheus instance.

func (*LocalStore) LabelNames added in v0.13.0

LabelNames returns all known label names.

func (*LocalStore) LabelValues added in v0.13.0

LabelValues returns all known label values for a given label name.

func (*LocalStore) Series added in v0.13.0

Series returns all series for a requested time range and label matcher. The returned data may exceed the requested time bounds.

type MultiTSDBStore added in v0.13.0

type MultiTSDBStore struct {
	// contains filtered or unexported fields
}

MultiTSDBStore implements the Store interface backed by multiple TSDBStore instances. TODO(bwplotka): Remove this and use Proxy instead. Details: https://github.com/thanos-io/thanos/issues/2864

func NewMultiTSDBStore added in v0.13.0

func NewMultiTSDBStore(logger log.Logger, _ prometheus.Registerer, component component.SourceStoreAPI, tsdbStores func() map[string]storepb.StoreServer) *MultiTSDBStore

NewMultiTSDBStore creates a new MultiTSDBStore.

func (*MultiTSDBStore) Info added in v0.13.0

Info returns store merged information about the underlying TSDBStore instances.

func (*MultiTSDBStore) LabelNames added in v0.13.0

LabelNames returns all known label names.

func (*MultiTSDBStore) LabelValues added in v0.13.0

LabelValues returns all known label values for a given label name.

func (*MultiTSDBStore) Series added in v0.13.0

Series returns all series for a requested time range and label matcher. The returned data may exceed the requested time bounds. The data returned may have been read and merged from multiple underlying TSDBStore instances.

type PrometheusStore

type PrometheusStore struct {
	// contains filtered or unexported fields
}

PrometheusStore implements the store node API on top of the Prometheus remote read API.

func NewPrometheusStore

func NewPrometheusStore(
	logger log.Logger,
	reg prometheus.Registerer,
	client *promclient.Client,
	baseURL *url.URL,
	component component.StoreAPI,
	externalLabels func() labels.Labels,
	timestamps func() (mint int64, maxt int64),
) (*PrometheusStore, error)

NewPrometheusStore returns a new PrometheusStore that uses the given HTTP client to talk to Prometheus. It attaches the provided external labels to all results.

func (*PrometheusStore) Info

Info returns store information about the Prometheus instance. NOTE(bwplotka): MaxTime & MinTime are not accurate nor adjusted dynamically. This is fine for now, but might be needed in future.

func (*PrometheusStore) LabelNames

LabelNames returns all known label names.

func (*PrometheusStore) LabelValues

LabelValues returns all known label values for a given label name.

func (*PrometheusStore) Series

Series returns all series for a requested time range and label matcher.

type ProxyStore

type ProxyStore struct {
	// contains filtered or unexported fields
}

ProxyStore implements the store API that proxies request to all given underlying stores.

func NewProxyStore

func NewProxyStore(
	logger log.Logger,
	reg prometheus.Registerer,
	stores func() []Client,
	component component.StoreAPI,
	selectorLabels labels.Labels,
	responseTimeout time.Duration,
) *ProxyStore

NewProxyStore returns a new ProxyStore that uses the given clients that implements storeAPI to fan-in all series to the client. Note that there is no deduplication support. Deduplication should be done on the highest level (just before PromQL).

func (*ProxyStore) Info

Info returns store information about the external labels this store have.

func (*ProxyStore) LabelNames

LabelNames returns all known label names.

func (*ProxyStore) LabelValues

LabelValues returns all known label values for a given label name.

func (*ProxyStore) Series

Series returns all series for a requested time range and label matcher. Requested series are taken from other stores and proxied to RPC client. NOTE: Resulted data are not trimmed exactly to min and max time range.

type ReadWriteTSDBStore added in v0.11.0

type ReadWriteTSDBStore struct {
	storepb.StoreServer
	storepb.WriteableStoreServer
}

ReadWriteTSDBStore is a TSDBStore that can also be written to.

type TSDBReader added in v0.15.0

type TSDBReader interface {
	storage.ChunkQueryable
	StartTime() (int64, error)
}

type TSDBStore

type TSDBStore struct {
	// contains filtered or unexported fields
}

TSDBStore implements the store API against a local TSDB instance. It attaches the provided external labels to all results. It only responds with raw data and does not support downsampling.

func NewTSDBStore

func NewTSDBStore(logger log.Logger, _ prometheus.Registerer, db TSDBReader, component component.StoreAPI, externalLabels labels.Labels) *TSDBStore

NewTSDBStore creates a new TSDBStore.

func (*TSDBStore) Info

Info returns store information about the Prometheus instance.

func (*TSDBStore) LabelNames

LabelNames returns all known label names.

func (*TSDBStore) LabelValues

LabelValues returns all known label values for a given label name.

func (*TSDBStore) Series

Series returns all series for a requested time range and label matcher. The returned data may exceed the requested time bounds.

Directories

Path Synopsis

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL