Documentation ¶
Overview ¶
Package block contains common functionality for interacting with TSDB blocks in the context of Thanos.
Index ¶
- Constants
- Variables
- func Delete(ctx context.Context, logger log.Logger, bkt objstore.Bucket, id ulid.ULID) error
- func Download(ctx context.Context, logger log.Logger, bucket objstore.Bucket, id ulid.ULID, ...) error
- func DownloadMeta(ctx context.Context, logger log.Logger, bkt objstore.Bucket, id ulid.ULID) (metadata.Meta, error)
- func GetSegmentFiles(blockDir string) []string
- func IgnoreCompleteOutsideChunk(mint, maxt int64, _, curr *chunks.Meta) (bool, error)
- func IgnoreDuplicateOutsideChunk(_, _ int64, last, curr *chunks.Meta) (bool, error)
- func IgnoreIssue347OutsideChunk(_, maxt int64, _, curr *chunks.Meta) (bool, error)
- func IsBlockDir(path string) (id ulid.ULID, ok bool)
- func MarkForDeletion(ctx context.Context, logger log.Logger, bkt objstore.Bucket, id ulid.ULID, ...) error
- func MarkForNoCompact(ctx context.Context, logger log.Logger, bkt objstore.Bucket, id ulid.ULID, ...) error
- func ParseRelabelConfig(contentYaml []byte, supportedActions map[relabel.Action]struct{}) ([]*relabel.Config, error)
- func Repair(logger log.Logger, dir string, id ulid.ULID, source metadata.SourceType, ...) (resid ulid.ULID, err error)
- func Upload(ctx context.Context, logger log.Logger, bkt objstore.Bucket, bdir string, ...) error
- func UploadPromBlock(ctx context.Context, logger log.Logger, bkt objstore.Bucket, bdir string, ...) error
- func VerifyIndex(logger log.Logger, fn string, minTime, maxTime int64) error
- type BaseFetcher
- type ConsistencyDelayMetaFilter
- type DeduplicateFilter
- type DiskWriter
- type FetcherMetrics
- type HealthStats
- type IgnoreDeletionMarkFilter
- type LabelShardedMetaFilter
- type MetaFetcher
- type MetadataFetcher
- type MetadataFilter
- type MetadataModifier
- type Node
- type Reader
- type ReplicaLabelRemover
- type SeriesWriter
- type TimePartitionMetaFilter
- type Writer
Constants ¶
const ( // MetaFilename is the known JSON filename for meta information. MetaFilename = "meta.json" // IndexFilename is the known index file for block index. IndexFilename = "index" // IndexHeaderFilename is the canonical name for binary index header file that stores essential information. IndexHeaderFilename = "index-header" // ChunksDirname is the known dir name for chunks with compressed samples. ChunksDirname = "chunks" // DebugMetas is a directory for debug meta files that happen in the past. Useful for debugging. DebugMetas = "debug/metas" )
const ( CorruptedMeta = "corrupted-meta-json" NoMeta = "no-meta-json" LoadedMeta = "loaded" FailedMeta = "failed" // Blocks that are marked for deletion can be loaded as well. This is done to make sure that we load blocks that are meant to be deleted, // but don't have a replacement block yet. MarkedForDeletionMeta = "marked-for-deletion" // MarkedForNoCompactionMeta is label for blocks which are loaded but also marked for no compaction. This label is also counted in `loaded` label metric. MarkedForNoCompactionMeta = "marked-for-no-compact" )
const BlockIDLabel = "__block_id"
Special label that will have an ULID of the meta.json being referenced to.
const FetcherConcurrency = 32
Variables ¶
var ( ErrorSyncMetaNotFound = errors.New("meta.json not found") ErrorSyncMetaCorrupted = errors.New("meta.json corrupted") )
Functions ¶
func Delete ¶
Delete removes directory that is meant to be block directory. NOTE: Always prefer this method for deleting blocks.
- We have to delete block's files in the certain order (meta.json first and deletion-mark.json last) to ensure we don't end up with malformed partial blocks. Thanos system handles well partial blocks only if they don't have meta.json. If meta.json is present Thanos assumes valid block.
- This avoids deleting empty dir (whole bucket) by mistake.
func Download ¶
func Download(ctx context.Context, logger log.Logger, bucket objstore.Bucket, id ulid.ULID, dst string) error
Download downloads directory that is mean to be block directory. If any of the files have a hash calculated in the meta file and it matches with what is in the destination path then we do not download it. We always re-download the meta file.
func DownloadMeta ¶
func DownloadMeta(ctx context.Context, logger log.Logger, bkt objstore.Bucket, id ulid.ULID) (metadata.Meta, error)
DownloadMeta downloads only meta file from bucket by block ID. TODO(bwplotka): Differentiate between network error & partial upload.
func GetSegmentFiles ¶ added in v0.17.0
GetSegmentFiles returns list of segment files for given block. Paths are relative to the chunks directory. In case of errors, nil is returned.
func MarkForDeletion ¶ added in v0.12.0
func MarkForDeletion(ctx context.Context, logger log.Logger, bkt objstore.Bucket, id ulid.ULID, details string, markedForDeletion prometheus.Counter) error
MarkForDeletion creates a file which stores information about when the block was marked for deletion.
func MarkForNoCompact ¶ added in v0.17.0
func MarkForNoCompact(ctx context.Context, logger log.Logger, bkt objstore.Bucket, id ulid.ULID, reason metadata.NoCompactReason, details string, markedForNoCompact prometheus.Counter) error
MarkForNoCompact creates a file which marks block to be not compacted.
func ParseRelabelConfig ¶ added in v0.15.0
func ParseRelabelConfig(contentYaml []byte, supportedActions map[relabel.Action]struct{}) ([]*relabel.Config, error)
ParseRelabelConfig parses relabel configuration. If supportedActions not specified, all relabel actions are valid.
func Repair ¶
func Repair(logger log.Logger, dir string, id ulid.ULID, source metadata.SourceType, ignoreChkFns ...ignoreFnType) (resid ulid.ULID, err error)
Repair open the block with given id in dir and creates a new one with fixed data. It: - removes out of order duplicates - all "complete" outsiders (they will not accessed anyway) - removes all near "complete" outside chunks introduced by https://github.com/prometheus/tsdb/issues/347. Fixable inconsistencies are resolved in the new block. TODO(bplotka): https://github.com/thanos-io/thanos/issues/378.
func Upload ¶
func Upload(ctx context.Context, logger log.Logger, bkt objstore.Bucket, bdir string, hf metadata.HashFunc) error
Upload uploads a TSDB block to the object storage. It verifies basic features of Thanos block.
func UploadPromBlock ¶ added in v0.20.0
func UploadPromBlock(ctx context.Context, logger log.Logger, bkt objstore.Bucket, bdir string, hf metadata.HashFunc) error
UploadPromBlock uploads a TSDB block to the object storage. It assumes the block is used in Prometheus so it doesn't check Thanos external labels.
Types ¶
type BaseFetcher ¶ added in v0.12.0
type BaseFetcher struct {
// contains filtered or unexported fields
}
BaseFetcher is a struct that synchronizes filtered metadata of all block in the object storage with the local state. Go-routine safe.
func NewBaseFetcher ¶ added in v0.12.0
func NewBaseFetcher(logger log.Logger, concurrency int, bkt objstore.InstrumentedBucketReader, dir string, reg prometheus.Registerer) (*BaseFetcher, error)
NewBaseFetcher constructs BaseFetcher.
func (*BaseFetcher) NewMetaFetcher ¶ added in v0.12.0
func (f *BaseFetcher) NewMetaFetcher(reg prometheus.Registerer, filters []MetadataFilter, modifiers []MetadataModifier, logTags ...interface{}) *MetaFetcher
NewMetaFetcher transforms BaseFetcher into actually usable *MetaFetcher.
type ConsistencyDelayMetaFilter ¶ added in v0.11.0
type ConsistencyDelayMetaFilter struct {
// contains filtered or unexported fields
}
ConsistencyDelayMetaFilter is a BaseFetcher filter that filters out blocks that are created before a specified consistency delay. Not go-routine safe.
func NewConsistencyDelayMetaFilter ¶ added in v0.11.0
func NewConsistencyDelayMetaFilter(logger log.Logger, consistencyDelay time.Duration, reg prometheus.Registerer) *ConsistencyDelayMetaFilter
NewConsistencyDelayMetaFilter creates ConsistencyDelayMetaFilter.
type DeduplicateFilter ¶ added in v0.11.0
type DeduplicateFilter struct {
// contains filtered or unexported fields
}
DeduplicateFilter is a BaseFetcher filter that filters out older blocks that have exactly the same data. Not go-routine safe.
func NewDeduplicateFilter ¶ added in v0.11.0
func NewDeduplicateFilter() *DeduplicateFilter
NewDeduplicateFilter creates DeduplicateFilter.
func (*DeduplicateFilter) DuplicateIDs ¶ added in v0.11.0
func (f *DeduplicateFilter) DuplicateIDs() []ulid.ULID
DuplicateIDs returns slice of block ids that are filtered out by DeduplicateFilter.
func (*DeduplicateFilter) Filter ¶ added in v0.11.0
func (f *DeduplicateFilter) Filter(_ context.Context, metas map[ulid.ULID]*metadata.Meta, synced *extprom.TxGaugeVec) error
Filter filters out duplicate blocks that can be formed from two or more overlapping blocks that fully submatches the source blocks of the older blocks.
type DiskWriter ¶ added in v0.18.0
type DiskWriter struct {
// contains filtered or unexported fields
}
func NewDiskWriter ¶ added in v0.18.0
NewDiskWriter allows to write single TSDB block to disk and returns statistics. Destination block directory has to exists.
func (*DiskWriter) Flush ¶ added in v0.18.0
func (d *DiskWriter) Flush() (_ tsdb.BlockStats, err error)
func (*DiskWriter) WriteChunks ¶ added in v0.18.0
type FetcherMetrics ¶ added in v0.19.0
type FetcherMetrics struct { Syncs prometheus.Counter SyncFailures prometheus.Counter SyncDuration prometheus.Histogram Synced *extprom.TxGaugeVec Modified *extprom.TxGaugeVec }
FetcherMetrics holds metrics tracked by the metadata fetcher. This struct and its fields are exported to allow depending projects (eg. Cortex) to implement their own custom metadata fetcher while tracking compatible metrics.
func NewFetcherMetrics ¶ added in v0.19.0
func NewFetcherMetrics(reg prometheus.Registerer, syncedExtraLabels, modifiedExtraLabels [][]string) *FetcherMetrics
func (*FetcherMetrics) ResetTx ¶ added in v0.19.0
func (s *FetcherMetrics) ResetTx()
ResetTx starts new transaction for metrics tracked by transaction GaugeVec.
func (*FetcherMetrics) Submit ¶ added in v0.19.0
func (s *FetcherMetrics) Submit()
Submit applies new values for metrics tracked by transaction GaugeVec.
type HealthStats ¶ added in v0.18.0
type HealthStats struct { // TotalSeries represents total number of series in block. TotalSeries int64 // OutOfOrderSeries represents number of series that have out of order chunks. OutOfOrderSeries int // OutOfOrderChunks represents number of chunks that are out of order (older time range is after younger one). OutOfOrderChunks int // DuplicatedChunks represents number of chunks with same time ranges within same series, potential duplicates. DuplicatedChunks int // OutsideChunks represents number of all chunks that are before or after time range specified in block meta. OutsideChunks int // CompleteOutsideChunks is subset of OutsideChunks that will be never accessed. They are completely out of time range specified in block meta. CompleteOutsideChunks int // Issue347OutsideChunks represents subset of OutsideChunks that are outsiders caused by https://github.com/prometheus/tsdb/issues/347 // and is something that Thanos handle. // // Specifically we mean here chunks with minTime == block.maxTime and maxTime > block.MaxTime. These are // are segregated into separate counters. These chunks are safe to be deleted, since they are duplicated across 2 blocks. Issue347OutsideChunks int // OutOfOrderLabels represents the number of postings that contained out // of order labels, a bug present in Prometheus 2.8.0 and below. OutOfOrderLabels int // Debug Statistics. SeriesMinLifeDuration time.Duration SeriesAvgLifeDuration time.Duration SeriesMaxLifeDuration time.Duration SeriesMinLifeDurationWithoutSingleSampleSeries time.Duration SeriesAvgLifeDurationWithoutSingleSampleSeries time.Duration SeriesMaxLifeDurationWithoutSingleSampleSeries time.Duration SeriesMinChunks int64 SeriesAvgChunks int64 SeriesMaxChunks int64 TotalChunks int64 ChunkMinDuration time.Duration ChunkAvgDuration time.Duration ChunkMaxDuration time.Duration ChunkMinSize int64 ChunkAvgSize int64 ChunkMaxSize int64 SingleSampleSeries int64 SingleSampleChunks int64 LabelNamesCount int64 MetricLabelValuesCount int64 }
func GatherIndexHealthStats ¶ added in v0.18.0
func GatherIndexHealthStats(logger log.Logger, fn string, minTime, maxTime int64) (stats HealthStats, err error)
GatherIndexHealthStats returns useful counters as well as outsider chunks (chunks outside of block time range) that helps to assess index health. It considers https://github.com/prometheus/tsdb/issues/347 as something that Thanos can handle. See HealthStats.Issue347OutsideChunks for details.
func (HealthStats) AnyErr ¶ added in v0.18.0
func (i HealthStats) AnyErr() error
AnyErr returns error if stats indicates any block issue.
func (HealthStats) CriticalErr ¶ added in v0.18.0
func (i HealthStats) CriticalErr() error
CriticalErr returns error if stats indicates critical block issue, that might solved only by manual repair procedure.
func (HealthStats) Issue347OutsideChunksErr ¶ added in v0.18.0
func (i HealthStats) Issue347OutsideChunksErr() error
Issue347OutsideChunksErr returns error if stats indicates issue347 block issue, that is repaired explicitly before compaction (on plan block).
func (HealthStats) OutOfOrderChunksErr ¶ added in v0.23.0
func (i HealthStats) OutOfOrderChunksErr() error
func (HealthStats) PrometheusIssue5372Err ¶ added in v0.18.0
func (i HealthStats) PrometheusIssue5372Err() error
PrometheusIssue5372Err returns an error if the HealthStats object indicates postings with out of order labels. This is corrected by Prometheus Issue #5372 and affects Prometheus versions 2.8.0 and below.
type IgnoreDeletionMarkFilter ¶ added in v0.12.0
type IgnoreDeletionMarkFilter struct {
// contains filtered or unexported fields
}
IgnoreDeletionMarkFilter is a filter that filters out the blocks that are marked for deletion after a given delay. The delay duration is to make sure that the replacement block can be fetched before we filter out the old block. Delay is not considered when computing DeletionMarkBlocks map. Not go-routine safe.
func NewIgnoreDeletionMarkFilter ¶ added in v0.12.0
func NewIgnoreDeletionMarkFilter(logger log.Logger, bkt objstore.InstrumentedBucketReader, delay time.Duration, concurrency int) *IgnoreDeletionMarkFilter
NewIgnoreDeletionMarkFilter creates IgnoreDeletionMarkFilter.
func (*IgnoreDeletionMarkFilter) DeletionMarkBlocks ¶ added in v0.12.0
func (f *IgnoreDeletionMarkFilter) DeletionMarkBlocks() map[ulid.ULID]*metadata.DeletionMark
DeletionMarkBlocks returns block ids that were marked for deletion.
func (*IgnoreDeletionMarkFilter) Filter ¶ added in v0.12.0
func (f *IgnoreDeletionMarkFilter) Filter(ctx context.Context, metas map[ulid.ULID]*metadata.Meta, synced *extprom.TxGaugeVec) error
Filter filters out blocks that are marked for deletion after a given delay. It also returns the blocks that can be deleted since they were uploaded delay duration before current time.
type LabelShardedMetaFilter ¶ added in v0.10.0
type LabelShardedMetaFilter struct {
// contains filtered or unexported fields
}
LabelShardedMetaFilter represents struct that allows sharding. Not go-routine safe.
func NewLabelShardedMetaFilter ¶ added in v0.10.0
func NewLabelShardedMetaFilter(relabelConfig []*relabel.Config) *LabelShardedMetaFilter
NewLabelShardedMetaFilter creates LabelShardedMetaFilter.
type MetaFetcher ¶ added in v0.10.0
type MetaFetcher struct {
// contains filtered or unexported fields
}
func NewMetaFetcher ¶ added in v0.10.0
func NewMetaFetcher(logger log.Logger, concurrency int, bkt objstore.InstrumentedBucketReader, dir string, reg prometheus.Registerer, filters []MetadataFilter, modifiers []MetadataModifier) (*MetaFetcher, error)
NewMetaFetcher returns meta fetcher.
func NewRawMetaFetcher ¶ added in v0.19.0
func NewRawMetaFetcher(logger log.Logger, bkt objstore.InstrumentedBucketReader) (*MetaFetcher, error)
NewRawMetaFetcher returns basic meta fetcher without proper handling for eventual consistent backends or partial uploads. NOTE: Not suitable to use in production.
func (*MetaFetcher) Fetch ¶ added in v0.10.0
func (f *MetaFetcher) Fetch(ctx context.Context) (metas map[ulid.ULID]*metadata.Meta, partial map[ulid.ULID]error, err error)
Fetch returns all block metas as well as partial blocks (blocks without or with corrupted meta file) from the bucket. It's caller responsibility to not change the returned metadata files. Maps can be modified.
Returned error indicates a failure in fetching metadata. Returned meta can be assumed as correct, with some blocks missing.
func (*MetaFetcher) UpdateOnChange ¶ added in v0.12.0
func (f *MetaFetcher) UpdateOnChange(listener func([]metadata.Meta, error))
UpdateOnChange allows to add listener that will be update on every change.
type MetadataFetcher ¶ added in v0.10.0
type MetadataFilter ¶ added in v0.12.0
type MetadataModifier ¶ added in v0.12.0
type Reader ¶ added in v0.18.0
type Reader interface { // Index returns an IndexReader over the block's data. Index() (tsdb.IndexReader, error) // Chunks returns a ChunkReader over the block's data. Chunks() (tsdb.ChunkReader, error) // Meta returns block metadata file. Meta() tsdb.BlockMeta }
Reader is like tsdb.BlockReader but without tombstones and size methods.
type ReplicaLabelRemover ¶ added in v0.12.0
type ReplicaLabelRemover struct {
// contains filtered or unexported fields
}
ReplicaLabelRemover is a BaseFetcher modifier modifies external labels of existing blocks, it removes given replica labels from the metadata of blocks that have it.
func NewReplicaLabelRemover ¶ added in v0.12.0
func NewReplicaLabelRemover(logger log.Logger, replicaLabels []string) *ReplicaLabelRemover
NewReplicaLabelRemover creates a ReplicaLabelRemover.
func (*ReplicaLabelRemover) Modify ¶ added in v0.12.0
func (r *ReplicaLabelRemover) Modify(_ context.Context, metas map[ulid.ULID]*metadata.Meta, modified *extprom.TxGaugeVec) error
Modify modifies external labels of existing blocks, it removes given replica labels from the metadata of blocks that have it.
type SeriesWriter ¶ added in v0.18.0
type SeriesWriter interface { tsdb.IndexWriter tsdb.ChunkWriter }
SeriesWriter is interface for writing series into one or multiple Blocks. Statistics has to be counted by implementation.
type TimePartitionMetaFilter ¶ added in v0.10.0
type TimePartitionMetaFilter struct {
// contains filtered or unexported fields
}
TimePartitionMetaFilter is a BaseFetcher filter that filters out blocks that are outside of specified time range. Not go-routine safe.
func NewTimePartitionMetaFilter ¶ added in v0.10.0
func NewTimePartitionMetaFilter(MinTime, MaxTime model.TimeOrDurationValue) *TimePartitionMetaFilter
NewTimePartitionMetaFilter creates TimePartitionMetaFilter.
type Writer ¶ added in v0.18.0
type Writer interface { SeriesWriter Flush() (tsdb.BlockStats, error) }
Writer is interface for creating block(s).