Documentation ¶
Overview ¶
Package tsdb implements a durable time series database.
Index ¶
- Constants
- Variables
- func DedupeEntries(a [][]byte) [][]byte
- func ErrDatabaseNotFound(name string) error
- func ErrMeasurementNotFound(name string) error
- func IsNumeric(c *influxql.Call) bool
- func IsRetryable(err error) bool
- func MapCount(input *MapInput) interface{}
- func MapCountDistinct(input *MapInput) interface{}
- func MapDistinct(input *MapInput) interface{}
- func MapEcho(input *MapInput) interface{}
- func MapFirst(input *MapInput, fieldName string) interface{}
- func MapLast(input *MapInput, fieldName string) interface{}
- func MapMax(input *MapInput, fieldName string) interface{}
- func MapMean(input *MapInput) interface{}
- func MapMin(input *MapInput, fieldName string) interface{}
- func MapRawQuery(input *MapInput) interface{}
- func MapSpread(input *MapInput) interface{}
- func MapStddev(input *MapInput) interface{}
- func MapSum(input *MapInput) interface{}
- func MapTopBottom(input *MapInput, limit int, fields []string, argCount int, callName string) interface{}
- func MarshalTags(tags map[string]string) []byte
- func MeasurementFromSeriesKey(key string) string
- func NewSnapshotWriter(meta []byte, store *Store) (*snapshot.Writer, error)
- func NopWriteToCloser(w io.WriterTo) interface{ ... }
- func ProcessAggregateDerivative(results [][]interface{}, isNonNegative bool, interval time.Duration) [][]interface{}
- func ReduceCountDistinct(values []interface{}) interface{}
- func ReduceDistinct(values []interface{}) interface{}
- func ReduceFirst(values []interface{}) interface{}
- func ReduceLast(values []interface{}) interface{}
- func ReduceMax(values []interface{}) interface{}
- func ReduceMean(values []interface{}) interface{}
- func ReduceMedian(values []interface{}) interface{}
- func ReduceMin(values []interface{}) interface{}
- func ReducePercentile(values []interface{}, percentile float64) interface{}
- func ReduceSpread(values []interface{}) interface{}
- func ReduceStddev(values []interface{}) interface{}
- func ReduceSum(values []interface{}) interface{}
- func ReduceTopBottom(values []interface{}, limit int, fields []string, callName string) interface{}
- func RegisterEngine(name string, fn NewEngineFunc)
- func RegisteredEngines() []string
- type AggregateExecutor
- type AggregateMapper
- type ByteSlices
- type Config
- type Cursor
- type CursorSet
- type CursorSets
- type DatabaseIndex
- func (s *DatabaseIndex) CreateMeasurementIndexIfNotExists(name string) *Measurement
- func (s *DatabaseIndex) CreateSeriesIndexIfNotExists(measurementName string, series *Series) *Series
- func (db *DatabaseIndex) DropMeasurement(name string)
- func (db *DatabaseIndex) DropSeries(keys []string)
- func (di *DatabaseIndex) ExpandSources(sources influxql.Sources) (influxql.Sources, error)
- func (db *DatabaseIndex) ExpandWildcards(stmt *influxql.SelectStatement) (*influxql.SelectStatement, error)
- func (d *DatabaseIndex) Measurement(name string) *Measurement
- func (d *DatabaseIndex) MeasurementSeriesCounts() (nMeasurements int, nSeries int)
- func (db *DatabaseIndex) Measurements() Measurements
- func (d *DatabaseIndex) MeasurementsByName(names []string) []*Measurement
- func (db *DatabaseIndex) RewriteSelectStatement(stmt *influxql.SelectStatement) (*influxql.SelectStatement, error)
- func (d *DatabaseIndex) Series(key string) *Series
- func (d *DatabaseIndex) SeriesN() int
- func (s *DatabaseIndex) TagsForSeries(key string) map[string]string
- type Engine
- type EngineFormat
- type EngineOptions
- type ErrAuthorize
- type Executor
- type Field
- type FieldCodec
- func (f *FieldCodec) DecodeByID(targetID uint8, b []byte) (interface{}, error)
- func (f *FieldCodec) DecodeByName(name string, b []byte) (interface{}, error)
- func (f *FieldCodec) DecodeFields(b []byte) (map[uint8]interface{}, error)
- func (f *FieldCodec) DecodeFieldsWithNames(b []byte) (map[string]interface{}, error)
- func (f *FieldCodec) EncodeFields(values map[string]interface{}) ([]byte, error)
- func (f *FieldCodec) FieldByName(name string) *Field
- func (f *FieldCodec) FieldIDByName(s string) (uint8, error)
- func (f *FieldCodec) Fields() (a []*Field)
- type FieldCreate
- type FilterExprs
- type InterfaceValues
- type IntoWriteRequest
- type MapInput
- type MapItem
- type Mapper
- type MapperOutput
- type MapperOutputJSON
- type MapperValue
- type MapperValueJSON
- type MapperValues
- type Measurement
- func (m *Measurement) AddSeries(s *Series) bool
- func (m *Measurement) DimensionTagSets(stmt *influxql.SelectStatement) ([]*influxql.TagSet, error)
- func (m *Measurement) DropSeries(seriesID uint64)
- func (m *Measurement) FieldNames() (a []string)
- func (m *Measurement) HasField(name string) bool
- func (m *Measurement) HasSeries() bool
- func (m *Measurement) HasTagKey(k string) bool
- func (m *Measurement) SelectFields(stmt *influxql.SelectStatement) []string
- func (m *Measurement) SelectTags(stmt *influxql.SelectStatement) []string
- func (m *Measurement) SeriesByID(id uint64) *Series
- func (m *Measurement) SeriesKeys() []string
- func (m *Measurement) SetFieldName(name string)
- func (m *Measurement) TagKeys() []string
- func (m *Measurement) TagSets(stmt *influxql.SelectStatement, dimensions []string) ([]*influxql.TagSet, error)
- func (m *Measurement) TagValues(key string) []string
- func (m *Measurement) ValidateGroupBy(stmt *influxql.SelectStatement) error
- func (m *Measurement) WhereFields(stmt *influxql.SelectStatement) []string
- type MeasurementFields
- type MeasurementTagKeys
- type Measurements
- func (a Measurements) Len() int
- func (a Measurements) Less(i, j int) bool
- func (a Measurements) SelectFields(stmt *influxql.SelectStatement) []string
- func (a Measurements) SelectTags(stmt *influxql.SelectStatement) []string
- func (a Measurements) Swap(i, j int)
- func (a Measurements) WhereFields(stmt *influxql.SelectStatement) []string
- type MeasurementsTagKeys
- type NewEngineFunc
- type NumberType
- type PointBatcher
- type PointBatcherStats
- type PositionPoint
- type PositionPoints
- type QueryExecutor
- func (q *QueryExecutor) Authorize(u *meta.UserInfo, query *influxql.Query, database string) error
- func (q *QueryExecutor) ExecuteQuery(query *influxql.Query, database string, chunkSize int) (<-chan *influxql.Result, error)
- func (q *QueryExecutor) PlanSelect(stmt *influxql.SelectStatement, chunkSize int) (Executor, error)
- func (q *QueryExecutor) PlanShowMeasurements(stmt *influxql.ShowMeasurementsStatement, database string, chunkSize int) (Executor, error)
- func (q *QueryExecutor) PlanShowTagKeys(stmt *influxql.ShowTagKeysStatement, database string, chunkSize int) (Executor, error)
- func (q *QueryExecutor) SetLogger(l *log.Logger)
- type RawExecutor
- type RawMapper
- type RawQueryDerivativeProcessor
- type SelectInfo
- type Series
- type SeriesCreate
- type SeriesIDs
- func (a SeriesIDs) Equals(other SeriesIDs) bool
- func (a SeriesIDs) Intersect(other SeriesIDs) SeriesIDs
- func (a SeriesIDs) Len() int
- func (a SeriesIDs) Less(i, j int) bool
- func (a SeriesIDs) Reject(other SeriesIDs) SeriesIDs
- func (a SeriesIDs) Swap(i, j int)
- func (a SeriesIDs) Union(other SeriesIDs) SeriesIDs
- type Shard
- func (s *Shard) Close() error
- func (s *Shard) DeleteMeasurement(name string, seriesKeys []string) error
- func (s *Shard) DeleteSeries(keys []string) error
- func (s *Shard) DiskSize() (int64, error)
- func (s *Shard) FieldCodec(measurementName string) *FieldCodec
- func (s *Shard) Open() error
- func (s *Shard) Path() string
- func (s *Shard) PerformMaintenance()
- func (s *Shard) ReadOnlyTx() (Tx, error)
- func (s *Shard) SeriesCount() (int, error)
- func (s *Shard) ValidateAggregateFieldsInStatement(measurementName string, stmt *influxql.SelectStatement) error
- func (s *Shard) WritePoints(points []models.Point) error
- func (s *Shard) WriteTo(w io.Writer) (int64, error)
- type ShowMeasurementsExecutor
- type ShowMeasurementsMapper
- func (m *ShowMeasurementsMapper) Close()
- func (m *ShowMeasurementsMapper) Fields() []string
- func (m *ShowMeasurementsMapper) NextChunk() (interface{}, error)
- func (m *ShowMeasurementsMapper) Open() error
- func (m *ShowMeasurementsMapper) SetRemote(remote Mapper)
- func (m *ShowMeasurementsMapper) TagSets() []string
- type ShowTagKeysExecutor
- type ShowTagKeysMapper
- type StatefulMapper
- type Store
- func (s *Store) Close() error
- func (s *Store) CreateMapper(shardID uint64, stmt influxql.Statement, chunkSize int) (Mapper, error)
- func (s *Store) CreateShard(database, retentionPolicy string, shardID uint64) error
- func (s *Store) DatabaseIndex(name string) *DatabaseIndex
- func (s *Store) DatabaseIndexN() int
- func (s *Store) Databases() []string
- func (s *Store) DeleteDatabase(name string, shardIDs []uint64) error
- func (s *Store) DeleteShard(shardID uint64) error
- func (s *Store) DiskSize() (int64, error)
- func (s *Store) Measurement(database, name string) *Measurement
- func (s *Store) Open() error
- func (s *Store) Path() string
- func (s *Store) Shard(id uint64) *Shard
- func (s *Store) ShardIDs() []uint64
- func (s *Store) ShardN() int
- func (s *Store) ValidateAggregateFieldsInStatement(shardID uint64, measurementName string, stmt *influxql.SelectStatement) error
- func (s *Store) WriteToShard(shardID uint64, points []models.Point) error
- type TagFilter
- type TagSetCursor
- type TagSetCursors
- type TagsCursor
- type Tx
- type UnmarshalFunc
Constants ¶
const ( // DefaultEngine is the default engine for new shards DefaultEngine = "bz1" // DefaultMaxWALSize is the default size of the WAL before it is flushed. DefaultMaxWALSize = 100 * 1024 * 1024 // 100MB // DefaultWALFlushInterval is the frequency the WAL will get flushed if // it doesn't reach its size threshold. DefaultWALFlushInterval = 10 * time.Minute // DefaultWALPartitionFlushDelay is the sleep time between WAL partition flushes. DefaultWALPartitionFlushDelay = 2 * time.Second // DefaultReadySeriesSize of 32KB specifies when a series is eligible to be flushed DefaultReadySeriesSize = 30 * 1024 // DefaultCompactionThreshold flush and compact a partition once this ratio of keys are over the flush size DefaultCompactionThreshold = 0.5 // DefaultMaxSeriesSize specifies the size at which a series will be forced to flush DefaultMaxSeriesSize = 1024 * 1024 // DefaultFlushColdInterval specifies how long after a partition has been cold // for writes that a full flush and compaction are forced DefaultFlushColdInterval = 5 * time.Second // DefaultParititionSizeThreshold specifies when a partition gets to this size in // memory, we should slow down writes until it gets a chance to compact. // This will force clients to get backpressure if they're writing too fast. We need // this because the WAL can take writes much faster than the index. So eventually // we'll need to create backpressure, otherwise we'll fill up the memory and die. // This number multiplied by the parition count is roughly the max possible memory // size for the in-memory WAL cache. DefaultPartitionSizeThreshold = 50 * 1024 * 1024 // 50MB // Default WAL settings for the TSM1 WAL DefaultFlushMemorySizeThreshold = 5 * 1024 * 1024 // 5MB DefaultMaxMemorySizeThreshold = 100 * 1024 * 1024 // 100MB DefaultIndexCompactionAge = time.Minute DefaultIndexMinCompactionInterval = time.Minute DefaultIndexMinCompactionFileCount = 5 DefaultIndexCompactionFullAge = 5 * time.Minute )
const ( // Return an error if the user is trying to select more than this number of points in a group by statement. // Most likely they specified a group by interval without time boundaries. MaxGroupByPoints = 100000 // Since time is always selected, the column count when selecting only a single other value will be 2 SelectColumnCountWithOneValue = 2 // IgnoredChunkSize is what gets passed into Mapper.Begin for aggregate queries as they don't chunk points out IgnoredChunkSize = 0 )
const EOF = int64(-1)
EOF represents a "not found" key returned by a Cursor.
const (
MaintenanceCheckInterval = time.Minute
)
Variables ¶
var ( // ErrInvalidQuery is returned when executing an unknown query type. ErrInvalidQuery = errors.New("invalid query") // ErrNotExecuted is returned when a statement is not executed in a query. // This can occur when a previous statement in the same query has errored. ErrNotExecuted = errors.New("not executed") )
var ( // ErrFieldOverflow is returned when too many fields are created on a measurement. ErrFieldOverflow = errors.New("field overflow") // ErrFieldTypeConflict is returned when a new field already exists with a different type. ErrFieldTypeConflict = errors.New("field type conflict") // ErrFieldNotFound is returned when a field cannot be found. ErrFieldNotFound = errors.New("field not found") // ErrFieldUnmappedID is returned when the system is presented, during decode, with a field ID // there is no mapping for. ErrFieldUnmappedID = errors.New("field ID not mapped") )
var ( ErrShardNotFound = fmt.Errorf("shard not found") ErrStoreClosed = fmt.Errorf("store is closed") )
var ( // ErrFormatNotFound is returned when no format can be determined from a path. ErrFormatNotFound = errors.New("format not found") )
Functions ¶
func DedupeEntries ¶ added in v0.9.3
DedupeEntries returns slices with unique keys (the first 8 bytes).
func ErrDatabaseNotFound ¶
func ErrMeasurementNotFound ¶
func IsRetryable ¶
IsRetryable returns true if this error is temporary and could be retried
func MapCount ¶
func MapCount(input *MapInput) interface{}
MapCount computes the number of values in an iterator.
func MapCountDistinct ¶
func MapCountDistinct(input *MapInput) interface{}
MapCountDistinct computes the unique count of values in an iterator.
func MapDistinct ¶
func MapDistinct(input *MapInput) interface{}
MapDistinct computes the unique values in an iterator.
func MapEcho ¶
func MapEcho(input *MapInput) interface{}
MapEcho emits the data points for each group by interval
func MapFirst ¶
MapFirst collects the values to pass to the reducer This function assumes time ordered input
func MapMean ¶
func MapMean(input *MapInput) interface{}
MapMean computes the count and sum of values in an iterator to be combined by the reducer.
func MapRawQuery ¶
func MapRawQuery(input *MapInput) interface{}
MapRawQuery is for queries without aggregates
func MapSpread ¶
func MapSpread(input *MapInput) interface{}
MapSpread collects the values to pass to the reducer
func MapStddev ¶
func MapStddev(input *MapInput) interface{}
MapStddev collects the values to pass to the reducer
func MapSum ¶
func MapSum(input *MapInput) interface{}
MapSum computes the summation of values in an iterator.
func MapTopBottom ¶
func MapTopBottom(input *MapInput, limit int, fields []string, argCount int, callName string) interface{}
MapTopBottom emits the top/bottom data points for each group by interval
func MarshalTags ¶ added in v0.9.3
used to convert the tag set to bytes for use as a lookup key
func MeasurementFromSeriesKey ¶ added in v0.9.3
func NewSnapshotWriter ¶
NewSnapshotWriter returns a new snapshot.Writer that will write metadata and the store's shards to an archive.
func NopWriteToCloser ¶
NopWriteToCloser returns an io.WriterTo that implements io.Closer.
func ProcessAggregateDerivative ¶ added in v0.9.3
func ProcessAggregateDerivative(results [][]interface{}, isNonNegative bool, interval time.Duration) [][]interface{}
ProcessAggregateDerivative returns the derivatives of an aggregate result set
func ReduceCountDistinct ¶
func ReduceCountDistinct(values []interface{}) interface{}
ReduceCountDistinct finds the unique counts of values.
func ReduceDistinct ¶
func ReduceDistinct(values []interface{}) interface{}
ReduceDistinct finds the unique values for each key.
func ReduceFirst ¶
func ReduceFirst(values []interface{}) interface{}
ReduceFirst computes the first of value.
func ReduceLast ¶
func ReduceLast(values []interface{}) interface{}
ReduceLast computes the last of value.
func ReduceMax ¶
func ReduceMax(values []interface{}) interface{}
ReduceMax computes the max of value.
func ReduceMean ¶
func ReduceMean(values []interface{}) interface{}
ReduceMean computes the mean of values for each key.
func ReduceMedian ¶
func ReduceMedian(values []interface{}) interface{}
ReduceMedian computes the median of values
func ReduceMin ¶
func ReduceMin(values []interface{}) interface{}
ReduceMin computes the min of value.
func ReducePercentile ¶
func ReducePercentile(values []interface{}, percentile float64) interface{}
ReducePercentile computes the percentile of values for each key.
func ReduceSpread ¶
func ReduceSpread(values []interface{}) interface{}
ReduceSpread computes the spread of values.
func ReduceStddev ¶
func ReduceStddev(values []interface{}) interface{}
ReduceStddev computes the stddev of values.
func ReduceSum ¶
func ReduceSum(values []interface{}) interface{}
ReduceSum computes the sum of values for each key.
func ReduceTopBottom ¶
ReduceTop computes the top values for each key. This function assumes that its inputs are in sorted ascending order.
func RegisterEngine ¶ added in v0.9.3
func RegisterEngine(name string, fn NewEngineFunc)
RegisterEngine registers a storage engine initializer by name.
func RegisteredEngines ¶
func RegisteredEngines() []string
RegisteredEngines returns the slice of currently registered engines.
Types ¶
type AggregateExecutor ¶ added in v0.9.2
type AggregateExecutor struct {
// contains filtered or unexported fields
}
AggregateExecutor represents a mapper for execute aggregate SELECT statements.
func NewAggregateExecutor ¶ added in v0.9.2
func NewAggregateExecutor(stmt *influxql.SelectStatement, mappers []Mapper) *AggregateExecutor
NewAggregateExecutor returns a new AggregateExecutor.
func (*AggregateExecutor) Execute ¶ added in v0.9.2
func (e *AggregateExecutor) Execute() <-chan *models.Row
Execute begins execution of the query and returns a channel to receive rows.
type AggregateMapper ¶
type AggregateMapper struct {
// contains filtered or unexported fields
}
AggregateMapper runs the map phase for aggregate SELECT queries.
func NewAggregateMapper ¶
func NewAggregateMapper(sh *Shard, stmt *influxql.SelectStatement) *AggregateMapper
NewAggregateMapper returns a new instance of AggregateMapper.
func (*AggregateMapper) Fields ¶
func (m *AggregateMapper) Fields() []string
Fields returns all SELECT fields.
func (*AggregateMapper) NextChunk ¶
func (m *AggregateMapper) NextChunk() (interface{}, error)
NextChunk returns the next interval of data. Tagsets are always processed in the same order as AvailTagsSets(). When there is no more data for any tagset nil is returned.
func (*AggregateMapper) Open ¶
func (m *AggregateMapper) Open() error
Open opens and initializes the mapper.
func (*AggregateMapper) TagSets ¶
func (m *AggregateMapper) TagSets() []string
TagSets returns the list of tag sets for which this mapper has data.
type ByteSlices ¶ added in v0.9.3
type ByteSlices [][]byte
func (ByteSlices) Len ¶ added in v0.9.3
func (a ByteSlices) Len() int
func (ByteSlices) Less ¶ added in v0.9.3
func (a ByteSlices) Less(i, j int) bool
func (ByteSlices) Swap ¶ added in v0.9.3
func (a ByteSlices) Swap(i, j int)
type Config ¶
type Config struct { Dir string `toml:"dir"` Engine string `toml:"engine"` // WAL config options for b1 (introduced in 0.9.2) MaxWALSize int `toml:"max-wal-size"` WALFlushInterval toml.Duration `toml:"wal-flush-interval"` WALPartitionFlushDelay toml.Duration `toml:"wal-partition-flush-delay"` // WAL configuration options for bz1 (introduced in 0.9.3) WALDir string `toml:"wal-dir"` WALLoggingEnabled bool `toml:"wal-logging-enabled"` WALReadySeriesSize int `toml:"wal-ready-series-size"` WALCompactionThreshold float64 `toml:"wal-compaction-threshold"` WALMaxSeriesSize int `toml:"wal-max-series-size"` WALFlushColdInterval toml.Duration `toml:"wal-flush-cold-interval"` WALPartitionSizeThreshold uint64 `toml:"wal-partition-size-threshold"` // WAL configuration options for tsm1 introduced in 0.9.5 WALFlushMemorySizeThreshold int `toml:"wal-flush-memory-size-threshold"` WALMaxMemorySizeThreshold int `toml:"wal-max-memory-size-threshold"` // IndexCompactionAge specifies the duration after the data file creation time // at which it is eligible to be compacted IndexCompactionAge time.Duration `toml:"index-compaction-age"` // IndexMinimumCompactionInterval specifies the minimum amount of time that must // pass after a compaction before another compaction is run IndexMinCompactionInterval time.Duration `toml:"index-min-compaction-interval"` // IndexCompactionFileCount specifies the minimum number of data files that // must be eligible for compaction before actually running one IndexMinCompactionFileCount int `toml:"index-compaction-min-file-count"` // IndexCompactionFullAge specifies how long after the last write was received // in the WAL that a full compaction should be performed. IndexCompactionFullAge time.Duration `toml:"index-compaction-full-age"` // Query logging QueryLogEnabled bool `toml:"query-log-enabled"` }
type Cursor ¶ added in v0.9.3
type Cursor interface { SeekTo(seek int64) (key int64, value interface{}) Next() (key int64, value interface{}) Ascending() bool }
Cursor represents an iterator over a series.
func MultiCursor ¶ added in v0.9.3
MultiCursor returns a single cursor that combines the results of all cursors in order.
If the same key is returned from multiple cursors then the first cursor specified will take precendence. A key will only be returned once from the returned cursor.
type CursorSet ¶
type CursorSet struct { Measurement string Tags map[string]string Key string Cursors []*TagsCursor }
type CursorSets ¶
type CursorSets []CursorSet
CursorSets represents a sortable slice of CursorSet.
func (CursorSets) Keys ¶
func (a CursorSets) Keys() []string
func (CursorSets) Len ¶
func (a CursorSets) Len() int
func (CursorSets) Less ¶
func (a CursorSets) Less(i, j int) bool
func (CursorSets) Swap ¶
func (a CursorSets) Swap(i, j int)
type DatabaseIndex ¶
type DatabaseIndex struct {
// contains filtered or unexported fields
}
DatabaseIndex is the in memory index of a collection of measurements, time series, and their tags. Exported functions are goroutine safe while un-exported functions assume the caller will use the appropriate locks
func NewDatabaseIndex ¶
func NewDatabaseIndex() *DatabaseIndex
func (*DatabaseIndex) CreateMeasurementIndexIfNotExists ¶ added in v0.9.3
func (s *DatabaseIndex) CreateMeasurementIndexIfNotExists(name string) *Measurement
CreateMeasurementIndexIfNotExists creates or retrieves an in memory index object for the measurement
func (*DatabaseIndex) CreateSeriesIndexIfNotExists ¶ added in v0.9.3
func (s *DatabaseIndex) CreateSeriesIndexIfNotExists(measurementName string, series *Series) *Series
CreateSeriesIndexIfNotExists adds the series for the given measurement to the index and sets its ID or returns the existing series object
func (*DatabaseIndex) DropMeasurement ¶
func (db *DatabaseIndex) DropMeasurement(name string)
DropMeasurement removes the measurement and all of its underlying series from the database index
func (*DatabaseIndex) DropSeries ¶
func (db *DatabaseIndex) DropSeries(keys []string)
DropSeries removes the series keys and their tags from the index
func (*DatabaseIndex) ExpandSources ¶
expandSources expands regex sources and removes duplicates. NOTE: sources must be normalized (db and rp set) before calling this function.
func (*DatabaseIndex) ExpandWildcards ¶
func (db *DatabaseIndex) ExpandWildcards(stmt *influxql.SelectStatement) (*influxql.SelectStatement, error)
expandWildcards returns a new SelectStatement with wildcards expanded If only a `SELECT *` is present, without a `GROUP BY *`, both tags and fields expand in the SELECT If a `SELECT *` and a `GROUP BY *` are both present, then only fiels are expanded in the `SELECT` and only tags are expanded in the `GROUP BY`
func (*DatabaseIndex) Measurement ¶
func (d *DatabaseIndex) Measurement(name string) *Measurement
Measurement returns the measurement object from the index by the name
func (*DatabaseIndex) MeasurementSeriesCounts ¶
func (d *DatabaseIndex) MeasurementSeriesCounts() (nMeasurements int, nSeries int)
MeasurementSeriesCounts returns the number of measurements and series currently indexed by the database. Useful for reporting and monitoring.
func (*DatabaseIndex) Measurements ¶
func (db *DatabaseIndex) Measurements() Measurements
Measurements returns a list of all measurements.
func (*DatabaseIndex) MeasurementsByName ¶
func (d *DatabaseIndex) MeasurementsByName(names []string) []*Measurement
MeasurementsByName returns a list of measurements.
func (*DatabaseIndex) RewriteSelectStatement ¶
func (db *DatabaseIndex) RewriteSelectStatement(stmt *influxql.SelectStatement) (*influxql.SelectStatement, error)
RewriteSelectStatement performs any necessary query re-writing.
func (*DatabaseIndex) Series ¶ added in v0.9.3
func (d *DatabaseIndex) Series(key string) *Series
Series returns a series by key.
func (*DatabaseIndex) SeriesN ¶ added in v0.9.3
func (d *DatabaseIndex) SeriesN() int
SeriesN returns the number of series.
func (*DatabaseIndex) TagsForSeries ¶ added in v0.9.3
func (s *DatabaseIndex) TagsForSeries(key string) map[string]string
TagsForSeries returns the tag map for the passed in series
type Engine ¶ added in v0.9.3
type Engine interface { Open() error Close() error SetLogOutput(io.Writer) LoadMetadataIndex(shard *Shard, index *DatabaseIndex, measurementFields map[string]*MeasurementFields) error Begin(writable bool) (Tx, error) WritePoints(points []models.Point, measurementFieldsToSave map[string]*MeasurementFields, seriesToCreate []*SeriesCreate) error DeleteSeries(keys []string) error DeleteMeasurement(name string, seriesKeys []string) error SeriesCount() (n int, err error) // PerformMaintenance will get called periodically by the store PerformMaintenance() // Format will return the format for the engine Format() EngineFormat io.WriterTo }
Engine represents a swappable storage engine for the shard.
type EngineFormat ¶
type EngineFormat int
const ( B1Format EngineFormat = iota BZ1Format TSM1Format )
type EngineOptions ¶ added in v0.9.3
type EngineOptions struct { EngineVersion string MaxWALSize int WALFlushInterval time.Duration WALPartitionFlushDelay time.Duration Config Config }
EngineOptions represents the options used to initialize the engine.
func NewEngineOptions ¶ added in v0.9.3
func NewEngineOptions() EngineOptions
NewEngineOptions returns the default options.
type ErrAuthorize ¶
type ErrAuthorize struct {
// contains filtered or unexported fields
}
ErrAuthorize represents an authorization error.
func NewErrAuthorize ¶ added in v0.9.2
func NewErrAuthorize(qe *QueryExecutor, q *influxql.Query, u, db, m string) *ErrAuthorize
newAuthorizationError returns a new instance of AuthorizationError.
func (ErrAuthorize) Error ¶
func (e ErrAuthorize) Error() string
Error returns the text of the error.
type Field ¶ added in v0.9.3
type Field struct { ID uint8 `json:"id,omitempty"` Name string `json:"name,omitempty"` Type influxql.DataType `json:"type,omitempty"` }
Field represents a series field.
type FieldCodec ¶
type FieldCodec struct {
// contains filtered or unexported fields
}
FieldCodec provides encoding and decoding functionality for the fields of a given Measurement. It is a distinct type to avoid locking writes on this node while potentially long-running queries are executing.
It is not affected by changes to the Measurement object after codec creation. TODO: this shouldn't be exported. nothing outside the shard should know about field encodings.
However, this is here until tx.go and the engine get refactored into tsdb.
func NewFieldCodec ¶ added in v0.9.3
func NewFieldCodec(fields map[string]*Field) *FieldCodec
NewFieldCodec returns a FieldCodec for the given Measurement. Must be called with a RLock that protects the Measurement.
func (*FieldCodec) DecodeByID ¶
func (f *FieldCodec) DecodeByID(targetID uint8, b []byte) (interface{}, error)
DecodeByID scans a byte slice for a field with the given ID, converts it to its expected type, and return that value. TODO: shouldn't be exported. refactor engine
func (*FieldCodec) DecodeByName ¶ added in v0.9.2
func (f *FieldCodec) DecodeByName(name string, b []byte) (interface{}, error)
DecodeByName scans a byte slice for a field with the given name, converts it to its expected type, and return that value.
func (*FieldCodec) DecodeFields ¶
func (f *FieldCodec) DecodeFields(b []byte) (map[uint8]interface{}, error)
DecodeFields decodes a byte slice into a set of field ids and values.
func (*FieldCodec) DecodeFieldsWithNames ¶
func (f *FieldCodec) DecodeFieldsWithNames(b []byte) (map[string]interface{}, error)
DecodeFieldsWithNames decodes a byte slice into a set of field names and values TODO: shouldn't be exported. refactor engine
func (*FieldCodec) EncodeFields ¶
func (f *FieldCodec) EncodeFields(values map[string]interface{}) ([]byte, error)
EncodeFields converts a map of values with string keys to a byte slice of field IDs and values.
If a field exists in the codec, but its type is different, an error is returned. If a field is not present in the codec, the system panics.
func (*FieldCodec) FieldByName ¶
func (f *FieldCodec) FieldByName(name string) *Field
FieldByName returns the field by its name. It will return a nil if not found
func (*FieldCodec) FieldIDByName ¶
func (f *FieldCodec) FieldIDByName(s string) (uint8, error)
TODO: this shouldn't be exported. remove when tx.go and engine.go get refactored into tsdb
func (*FieldCodec) Fields ¶
func (f *FieldCodec) Fields() (a []*Field)
type FieldCreate ¶ added in v0.9.3
struct to hold information for a field to create on a measurement
type FilterExprs ¶
FilterExprs represents a map of series IDs to filter expressions.
func (FilterExprs) DeleteBoolLiteralTrues ¶
func (fe FilterExprs) DeleteBoolLiteralTrues()
DeleteBoolLiteralTrues deletes all elements whose filter expression is a boolean literal true.
type InterfaceValues ¶
type InterfaceValues []interface{}
func (InterfaceValues) Len ¶
func (d InterfaceValues) Len() int
func (InterfaceValues) Less ¶
func (d InterfaceValues) Less(i, j int) bool
func (InterfaceValues) Swap ¶
func (d InterfaceValues) Swap(i, j int)
type IntoWriteRequest ¶
partial copy of cluster.WriteRequest
type MapItem ¶
type MapItem struct { Timestamp int64 Value interface{} // TODO(benbjohnson): // Move fields and tags up to MapInput. Currently the engine combines // multiple series together during processing. This needs to be fixed so // that each map function only operates on a single series at a time instead. Fields map[string]interface{} Tags map[string]string }
MapItem represents a single item in a collection that's processed by the mapper.
type Mapper ¶ added in v0.9.2
type Mapper interface { Open() error TagSets() []string Fields() []string NextChunk() (interface{}, error) Close() }
Mapper is the interface all Mapper types must implement.
type MapperOutput ¶ added in v0.9.3
type MapperOutput struct { Name string `json:"name,omitempty"` Tags map[string]string `json:"tags,omitempty"` Fields []string `json:"fields,omitempty"` // Field names of returned data. Values []*MapperValue `json:"values,omitempty"` // For aggregates contains a single value at [0] // contains filtered or unexported fields }
func (*MapperOutput) MarshalJSON ¶
func (mo *MapperOutput) MarshalJSON() ([]byte, error)
MarshalJSON returns the JSON-encoded representation of a MapperOutput.
type MapperOutputJSON ¶
type MapperOutputJSON struct { Name string `json:"name,omitempty"` Tags map[string]string `json:"tags,omitempty"` Fields []string `json:"fields,omitempty"` // Field names of returned data. Values json.RawMessage `json:"values,omitempty"` }
MapperOutputJSON is the JSON-encoded representation of MapperOutput. The query data is represented as a raw JSON message, so decode is delayed, and can proceed in a custom manner.
type MapperValue ¶ added in v0.9.3
type MapperValue struct { Time int64 `json:"time,omitempty"` // Ignored for aggregate output. Value interface{} `json:"value,omitempty"` // For aggregate, contains interval time multiple values. Tags map[string]string `json:"tags,omitempty"` // Meta tags for results }
MapperValue is a complex type, which can encapsulate data from both raw and aggregate mappers. This currently allows marshalling and network system to remain simpler. For aggregate output Time is ignored, and actual Time-Value pairs are contained soley within the Value field.
func (*MapperValue) MarshalJSON ¶
func (mv *MapperValue) MarshalJSON() ([]byte, error)
MarshalJSON returns the JSON-encoded representation of a MapperValue.
type MapperValueJSON ¶
type MapperValueJSON struct { Time int64 `json:"time,omitempty"` RawData []byte `json:"rdata,omitempty"` AggData [][]byte `json:"adata,omitempty"` Tags map[string]string `json:"tags,omitempty"` }
MapperValueJSON is the JSON-encoded representation of MapperValue. Because MapperValue is a complex type, custom JSON encoding is required so that none of the types contained within a MapperValue are "lost", and so the data are encoded as byte slices where necessary.
type MapperValues ¶ added in v0.9.3
type MapperValues []*MapperValue
func (MapperValues) Len ¶ added in v0.9.3
func (a MapperValues) Len() int
func (MapperValues) Less ¶ added in v0.9.3
func (a MapperValues) Less(i, j int) bool
func (MapperValues) Swap ¶ added in v0.9.3
func (a MapperValues) Swap(i, j int)
type Measurement ¶
type Measurement struct { Name string `json:"name,omitempty"` // contains filtered or unexported fields }
Measurement represents a collection of time series in a database. It also contains in memory structures for indexing tags. Exported functions are goroutine safe while un-exported functions assume the caller will use the appropriate locks
func NewMeasurement ¶
func NewMeasurement(name string, idx *DatabaseIndex) *Measurement
NewMeasurement allocates and initializes a new Measurement.
func (*Measurement) AddSeries ¶
func (m *Measurement) AddSeries(s *Series) bool
AddSeries will add a series to the measurementIndex. Returns false if already present
func (*Measurement) DimensionTagSets ¶
func (m *Measurement) DimensionTagSets(stmt *influxql.SelectStatement) ([]*influxql.TagSet, error)
DimensionTagSets returns list of tag sets from the GROUP BY section of stmt.
func (*Measurement) DropSeries ¶
func (m *Measurement) DropSeries(seriesID uint64)
DropSeries will remove a series from the measurementIndex.
func (*Measurement) FieldNames ¶
func (m *Measurement) FieldNames() (a []string)
FieldNames returns a list of the measurement's field names
func (*Measurement) HasField ¶
func (m *Measurement) HasField(name string) bool
HasField returns true if the measurement has a field by the given name
func (*Measurement) HasSeries ¶
func (m *Measurement) HasSeries() bool
HasSeries returns true if there is at least 1 series under this measurement
func (*Measurement) HasTagKey ¶
func (m *Measurement) HasTagKey(k string) bool
HasTagKey returns true if at least one series in this measurement has written a value for the passed in tag key
func (*Measurement) SelectFields ¶
func (m *Measurement) SelectFields(stmt *influxql.SelectStatement) []string
SelectFields returns a list of fields in the SELECT section of stmt.
func (*Measurement) SelectTags ¶
func (m *Measurement) SelectTags(stmt *influxql.SelectStatement) []string
SelectTags returns a list of non-field tags in the SELECT section of stmt.
func (*Measurement) SeriesByID ¶ added in v0.9.3
func (m *Measurement) SeriesByID(id uint64) *Series
SeriesByID returns a series by identifier.
func (*Measurement) SeriesKeys ¶
func (m *Measurement) SeriesKeys() []string
SeriesKeys returns the keys of every series in this measurement
func (*Measurement) SetFieldName ¶ added in v0.9.3
func (m *Measurement) SetFieldName(name string)
SetFieldName adds the field name to the measurement.
func (*Measurement) TagKeys ¶
func (m *Measurement) TagKeys() []string
TagKeys returns a list of the measurement's tag names.
func (*Measurement) TagSets ¶
func (m *Measurement) TagSets(stmt *influxql.SelectStatement, dimensions []string) ([]*influxql.TagSet, error)
tagSets returns the unique tag sets that exist for the given tag keys. This is used to determine what composite series will be created by a group by. i.e. "group by region" should return: {"region":"uswest"}, {"region":"useast"} or region, service returns {"region": "uswest", "service": "redis"}, {"region": "uswest", "service": "mysql"}, etc... This will also populate the TagSet objects with the series IDs that match each tagset and any influx filter expression that goes with the series TODO: this shouldn't be exported. However, until tx.go and the engine get refactored into tsdb, we need it.
func (*Measurement) TagValues ¶ added in v0.9.4
func (m *Measurement) TagValues(key string) []string
TagValues returns all the values for the given tag key
func (*Measurement) ValidateGroupBy ¶ added in v0.9.2
func (m *Measurement) ValidateGroupBy(stmt *influxql.SelectStatement) error
ValidateGroupBy ensures that the GROUP BY is not a field.
func (*Measurement) WhereFields ¶
func (m *Measurement) WhereFields(stmt *influxql.SelectStatement) []string
WhereFields returns a list of non-"time" fields in the WHERE section of stmt.
type MeasurementFields ¶ added in v0.9.3
type MeasurementFields struct { Fields map[string]*Field `json:"fields"` Codec *FieldCodec }
func (*MeasurementFields) CreateFieldIfNotExists ¶ added in v0.9.3
func (m *MeasurementFields) CreateFieldIfNotExists(name string, typ influxql.DataType, limitCount bool) error
CreateFieldIfNotExists creates a new field with an autoincrementing ID. Returns an error if 255 fields have already been created on the measurement or the fields already exists with a different type.
func (*MeasurementFields) MarshalBinary ¶ added in v0.9.3
func (m *MeasurementFields) MarshalBinary() ([]byte, error)
MarshalBinary encodes the object to a binary format.
func (*MeasurementFields) UnmarshalBinary ¶ added in v0.9.3
func (m *MeasurementFields) UnmarshalBinary(buf []byte) error
UnmarshalBinary decodes the object from a binary format.
type MeasurementTagKeys ¶
type MeasurementTagKeys struct { Measurement string `json:"measurement"` TagKeys []string `json:"tagkeys"` }
MeasurementTagKeys represents measurement tag keys.
type Measurements ¶
type Measurements []*Measurement
Measurements represents a list of *Measurement.
func (Measurements) Len ¶
func (a Measurements) Len() int
func (Measurements) Less ¶
func (a Measurements) Less(i, j int) bool
func (Measurements) SelectFields ¶
func (a Measurements) SelectFields(stmt *influxql.SelectStatement) []string
SelectFields returns a list of fields in the SELECT section of stmt.
func (Measurements) SelectTags ¶
func (a Measurements) SelectTags(stmt *influxql.SelectStatement) []string
SelectTags returns a list of non-field tags in the SELECT section of stmt.
func (Measurements) Swap ¶
func (a Measurements) Swap(i, j int)
func (Measurements) WhereFields ¶
func (a Measurements) WhereFields(stmt *influxql.SelectStatement) []string
WhereFields returns a list of non-"time" fields in the WHERE section of stmt.
type MeasurementsTagKeys ¶
type MeasurementsTagKeys []*MeasurementTagKeys
MeasurementsTagKeys represents tag keys for multiple measurements.
func (MeasurementsTagKeys) Len ¶
func (a MeasurementsTagKeys) Len() int
func (MeasurementsTagKeys) Less ¶
func (a MeasurementsTagKeys) Less(i, j int) bool
func (MeasurementsTagKeys) Size ¶
func (a MeasurementsTagKeys) Size() int
Size returns the total string length of measurement names & tag keys.
func (MeasurementsTagKeys) Swap ¶
func (a MeasurementsTagKeys) Swap(i, j int)
type NewEngineFunc ¶ added in v0.9.3
type NewEngineFunc func(path string, walPath string, options EngineOptions) Engine
NewEngineFunc creates a new engine.
type PointBatcher ¶
type PointBatcher struct {
// contains filtered or unexported fields
}
PointBatcher accepts Points and will emit a batch of those points when either a) the batch reaches a certain size, or b) a certain time passes.
func NewPointBatcher ¶
func NewPointBatcher(sz int, bp int, d time.Duration) *PointBatcher
NewPointBatcher returns a new PointBatcher. sz is the batching size, bp is the maximum number of batches that may be pending. d is the time after which a batch will be emitted after the first point is received for the batch, regardless of its size.
func (*PointBatcher) Flush ¶
func (b *PointBatcher) Flush()
Flush instructs the batcher to emit any pending points in a batch, regardless of batch size. If there are no pending points, no batch is emitted.
func (*PointBatcher) In ¶
func (b *PointBatcher) In() chan<- models.Point
In returns the channel to which points should be written.
func (*PointBatcher) Out ¶
func (b *PointBatcher) Out() <-chan []models.Point
Out returns the channel from which batches should be read.
func (*PointBatcher) Start ¶
func (b *PointBatcher) Start()
Start starts the batching process. Returns the in and out channels for points and point-batches respectively.
func (*PointBatcher) Stats ¶
func (b *PointBatcher) Stats() *PointBatcherStats
Stats returns a PointBatcherStats object for the PointBatcher. While the each statistic should be closely correlated with each other statistic, it is not guaranteed.
func (*PointBatcher) Stop ¶
func (b *PointBatcher) Stop()
type PointBatcherStats ¶
type PointBatcherStats struct { BatchTotal uint64 // Total count of batches transmitted. PointTotal uint64 // Total count of points processed. SizeTotal uint64 // Number of batches that reached size threshold. TimeoutTotal uint64 // Number of timeouts that occurred. }
PointBatcherStats are the statistics each batcher tracks.
type PositionPoint ¶
type PositionPoint struct { Time int64 Value interface{} Fields map[string]interface{} Tags map[string]string }
PositionPoint will return all data points from a written point that were selected in the query to be used in the post processing phase of the query executor to fill in additional tag and field values
type PositionPoints ¶
type PositionPoints []PositionPoint
PositionPoints is a slice of PositionPoints used to return richer data from a reduce func
type QueryExecutor ¶
type QueryExecutor struct { // The meta store for accessing and updating cluster and schema data. MetaStore interface { Database(name string) (*meta.DatabaseInfo, error) Databases() ([]meta.DatabaseInfo, error) User(name string) (*meta.UserInfo, error) AdminUserExists() (bool, error) Authenticate(username, password string) (*meta.UserInfo, error) RetentionPolicy(database, name string) (rpi *meta.RetentionPolicyInfo, err error) UserCount() (int, error) ShardGroupsByTimeRange(database, policy string, min, max time.Time) (a []meta.ShardGroupInfo, err error) NodeID() uint64 } // Executes statements relating to meta data. MetaStatementExecutor interface { ExecuteStatement(stmt influxql.Statement) *influxql.Result } // Execute statements relating to statistics and diagnostics. MonitorStatementExecutor interface { ExecuteStatement(stmt influxql.Statement) *influxql.Result } // Maps shards for queries. ShardMapper interface { CreateMapper(shard meta.ShardInfo, stmt influxql.Statement, chunkSize int) (Mapper, error) } IntoWriter interface { WritePointsInto(p *IntoWriteRequest) error } Logger *log.Logger QueryLogEnabled bool // the local data store Store *Store }
QueryExecutor executes every statement in an influxdb Query. It is responsible for coordinating between the local tsdb.Store, the meta.Store, and the other nodes in the cluster to run the query against their local tsdb.Stores. There should be one executor in a running process
func NewQueryExecutor ¶
func NewQueryExecutor(store *Store) *QueryExecutor
NewQueryExecutor returns an initialized QueryExecutor
func (*QueryExecutor) Authorize ¶
Authorize user u to execute query q on database. database can be "" for queries that do not require a database. If no user is provided it will return an error unless the query's first statement is to create a root user.
func (*QueryExecutor) ExecuteQuery ¶
func (q *QueryExecutor) ExecuteQuery(query *influxql.Query, database string, chunkSize int) (<-chan *influxql.Result, error)
ExecuteQuery executes an InfluxQL query against the server. It sends results down the passed in chan and closes it when done. It will close the chan on the first statement that throws an error.
func (*QueryExecutor) PlanSelect ¶ added in v0.9.4
func (q *QueryExecutor) PlanSelect(stmt *influxql.SelectStatement, chunkSize int) (Executor, error)
Plan creates an execution plan for the given SelectStatement and returns an Executor.
func (*QueryExecutor) PlanShowMeasurements ¶ added in v0.9.4
func (q *QueryExecutor) PlanShowMeasurements(stmt *influxql.ShowMeasurementsStatement, database string, chunkSize int) (Executor, error)
PlanShowMeasurements creates an execution plan for a SHOW TAG KEYS statement and returns an Executor.
func (*QueryExecutor) PlanShowTagKeys ¶
func (q *QueryExecutor) PlanShowTagKeys(stmt *influxql.ShowTagKeysStatement, database string, chunkSize int) (Executor, error)
PlanShowTagKeys creates an execution plan for a SHOW MEASUREMENTS statement and returns an Executor.
func (*QueryExecutor) SetLogger ¶ added in v0.9.4
func (q *QueryExecutor) SetLogger(l *log.Logger)
SetLogger sets the internal logger to the logger passed in.
type RawExecutor ¶ added in v0.9.2
type RawExecutor struct {
// contains filtered or unexported fields
}
func NewRawExecutor ¶ added in v0.9.2
func NewRawExecutor(stmt *influxql.SelectStatement, mappers []Mapper, chunkSize int) *RawExecutor
NewRawExecutor returns a new RawExecutor.
func (*RawExecutor) Execute ¶ added in v0.9.2
func (e *RawExecutor) Execute() <-chan *models.Row
Execute begins execution of the query and returns a channel to receive rows.
type RawMapper ¶ added in v0.9.2
type RawMapper struct { ChunkSize int // contains filtered or unexported fields }
RawMapper runs the map phase for non-aggregate, raw SELECT queries.
func NewRawMapper ¶ added in v0.9.2
func NewRawMapper(sh *Shard, stmt *influxql.SelectStatement) *RawMapper
NewRawMapper returns a new instance of RawMapper.
func (*RawMapper) NextChunk ¶ added in v0.9.2
NextChunk returns the next chunk of data. Data is ordered the same as TagSets. Each chunk contains one tag set. If there is no more data for any tagset, nil will be returned.
type RawQueryDerivativeProcessor ¶ added in v0.9.3
type RawQueryDerivativeProcessor struct { LastValueFromPreviousChunk *MapperValue IsNonNegative bool // Whether to drop negative differences DerivativeInterval time.Duration }
func (*RawQueryDerivativeProcessor) Process ¶ added in v0.9.3
func (rqdp *RawQueryDerivativeProcessor) Process(input []*MapperValue) []*MapperValue
type SelectInfo ¶
type Series ¶
Series belong to a Measurement and represent unique time series in a database
func (*Series) InitializeShards ¶ added in v0.9.3
func (s *Series) InitializeShards()
func (*Series) MarshalBinary ¶
MarshalBinary encodes the object to a binary format.
func (*Series) UnmarshalBinary ¶
UnmarshalBinary decodes the object from a binary format.
type SeriesCreate ¶ added in v0.9.3
struct to hold information for a series to create
type SeriesIDs ¶ added in v0.9.3
type SeriesIDs []uint64
SeriesIDs is a convenience type for sorting, checking equality, and doing union and intersection of collections of series ids.
func (SeriesIDs) Intersect ¶ added in v0.9.3
Intersect returns a new collection of series ids in sorted order that is the intersection of the two. The two collections must already be sorted.
type Shard ¶
type Shard struct { // The writer used by the logger. LogOutput io.Writer // contains filtered or unexported fields }
Shard represents a self-contained time series database. An inverted index of the measurement and tag data is kept along with the raw time series data. Data can be split across many shards. The query engine in TSDB is responsible for combining the output of many shards into a single query result.
func NewShard ¶
func NewShard(id uint64, index *DatabaseIndex, path string, walPath string, options EngineOptions) *Shard
NewShard returns a new initialized Shard. walPath doesn't apply to the b1 type index
func (*Shard) DeleteMeasurement ¶ added in v0.9.3
DeleteMeasurement deletes a measurement and all underlying series.
func (*Shard) DeleteSeries ¶ added in v0.9.3
DeleteSeries deletes a list of series.
func (*Shard) FieldCodec ¶
func (s *Shard) FieldCodec(measurementName string) *FieldCodec
TODO: this is temporarily exported to make tx.go work. When the query engine gets refactored into the tsdb package this should be removed. No one outside tsdb should know the underlying field encoding scheme.
func (*Shard) PerformMaintenance ¶
func (s *Shard) PerformMaintenance()
PerformMaintenance gets called periodically to have the engine perform any maintenance tasks like WAL flushing and compaction
func (*Shard) ReadOnlyTx ¶ added in v0.9.4
ReadOnlyTx returns a read-only transaction for the shard. The transaction must be rolled back to release resources.
func (*Shard) SeriesCount ¶ added in v0.9.1
SeriesCount returns the number of series buckets on the shard.
func (*Shard) ValidateAggregateFieldsInStatement ¶
func (s *Shard) ValidateAggregateFieldsInStatement(measurementName string, stmt *influxql.SelectStatement) error
func (*Shard) WritePoints ¶
WritePoints will write the raw data points and any new metadata to the index in the shard
type ShowMeasurementsExecutor ¶ added in v0.9.4
type ShowMeasurementsExecutor struct {
// contains filtered or unexported fields
}
ShowMeasurementsExecutor implements the Executor interface for a SHOW MEASUREMENTS statement.
func NewShowMeasurementsExecutor ¶ added in v0.9.4
func NewShowMeasurementsExecutor(stmt *influxql.ShowMeasurementsStatement, mappers []Mapper, chunkSize int) *ShowMeasurementsExecutor
NewShowMeasurementsExecutor returns a new ShowMeasurementsExecutor.
func (*ShowMeasurementsExecutor) Execute ¶ added in v0.9.4
func (e *ShowMeasurementsExecutor) Execute() <-chan *models.Row
Execute begins execution of the query and returns a channel to receive rows.
type ShowMeasurementsMapper ¶ added in v0.9.4
type ShowMeasurementsMapper struct { ChunkSize int // contains filtered or unexported fields }
ShowMeasurementsMapper is a mapper for collecting measurement names from a shard.
func NewShowMeasurementsMapper ¶ added in v0.9.4
func NewShowMeasurementsMapper(shard *Shard, stmt *influxql.ShowMeasurementsStatement) *ShowMeasurementsMapper
NewShowMeasurementsMapper returns a mapper for the given shard, which will return data for the meta statement.
func (*ShowMeasurementsMapper) Close ¶ added in v0.9.4
func (m *ShowMeasurementsMapper) Close()
Close closes the mapper.
func (*ShowMeasurementsMapper) Fields ¶ added in v0.9.4
func (m *ShowMeasurementsMapper) Fields() []string
Fields returns a list of field names for this mapper.
func (*ShowMeasurementsMapper) NextChunk ¶ added in v0.9.4
func (m *ShowMeasurementsMapper) NextChunk() (interface{}, error)
NextChunk returns the next chunk of measurement names.
func (*ShowMeasurementsMapper) Open ¶ added in v0.9.4
func (m *ShowMeasurementsMapper) Open() error
Open opens the mapper for use.
func (*ShowMeasurementsMapper) SetRemote ¶ added in v0.9.4
func (m *ShowMeasurementsMapper) SetRemote(remote Mapper)
SetRemote sets the remote mapper to use.
func (*ShowMeasurementsMapper) TagSets ¶ added in v0.9.4
func (m *ShowMeasurementsMapper) TagSets() []string
TagSets is only implemented on this mapper to satisfy the Mapper interface.
type ShowTagKeysExecutor ¶
type ShowTagKeysExecutor struct {
// contains filtered or unexported fields
}
ShowTagKeysExecutor implements the Executor interface for a SHOW MEASUREMENTS statement.
func NewShowTagKeysExecutor ¶
func NewShowTagKeysExecutor(stmt *influxql.ShowTagKeysStatement, mappers []Mapper, chunkSize int) *ShowTagKeysExecutor
NewShowTagKeysExecutor returns a new ShowTagKeysExecutor.
func (*ShowTagKeysExecutor) Execute ¶
func (e *ShowTagKeysExecutor) Execute() <-chan *models.Row
Execute begins execution of the query and returns a channel to receive rows.
type ShowTagKeysMapper ¶
type ShowTagKeysMapper struct {
// contains filtered or unexported fields
}
ShowTagKeysMapper is a mapper for collecting measurement names from a shard.
func NewShowTagKeysMapper ¶
func NewShowTagKeysMapper(shard *Shard, stmt *influxql.ShowTagKeysStatement, chunkSize int) *ShowTagKeysMapper
NewShowTagKeysMapper returns a mapper for the given shard, which will return data for the meta statement.
func (*ShowTagKeysMapper) Fields ¶
func (m *ShowTagKeysMapper) Fields() []string
Fields returns a list of field names for this mapper.
func (*ShowTagKeysMapper) NextChunk ¶
func (m *ShowTagKeysMapper) NextChunk() (interface{}, error)
NextChunk returns the next chunk of measurements and tag keys.
func (*ShowTagKeysMapper) Open ¶
func (m *ShowTagKeysMapper) Open() error
Open opens the mapper for use.
func (*ShowTagKeysMapper) SetRemote ¶
func (m *ShowTagKeysMapper) SetRemote(remote Mapper) error
SetRemote sets the remote mapper to use.
func (*ShowTagKeysMapper) TagSets ¶
func (m *ShowTagKeysMapper) TagSets() []string
TagSets is only implemented on this mapper to satisfy the Mapper interface.
type StatefulMapper ¶ added in v0.9.2
type StatefulMapper struct { Mapper // contains filtered or unexported fields }
StatefulMapper encapsulates a Mapper and some state that the executor needs to track for that mapper.
func (*StatefulMapper) NextChunk ¶ added in v0.9.2
func (sm *StatefulMapper) NextChunk() (*MapperOutput, error)
NextChunk wraps a RawMapper and some state.
type Store ¶
type Store struct { EngineOptions EngineOptions Logger *log.Logger // contains filtered or unexported fields }
func (*Store) CreateMapper ¶ added in v0.9.2
func (*Store) CreateShard ¶
func (*Store) DatabaseIndex ¶
func (s *Store) DatabaseIndex(name string) *DatabaseIndex
func (*Store) DatabaseIndexN ¶ added in v0.9.3
DatabaseIndexN returns the number of databases indicies in the store.
func (*Store) DeleteDatabase ¶
DeleteDatabase will close all shards associated with a database and remove the directory and files from disk.
func (*Store) DeleteShard ¶
DeleteShard removes a shard from disk.
func (*Store) DiskSize ¶ added in v0.9.4
DiskSize returns the size of all the shard files in bytes. This size does not include the WAL size.
func (*Store) Measurement ¶
func (s *Store) Measurement(database, name string) *Measurement
func (*Store) ValidateAggregateFieldsInStatement ¶
type TagSetCursor ¶
type TagSetCursor struct { SelectFields []string // fields to be selected // contains filtered or unexported fields }
TagSetCursor is virtual cursor that iterates over multiple TagsCursors.
func NewTagSetCursor ¶
func NewTagSetCursor(m string, t map[string]string, c []*TagsCursor) *TagSetCursor
NewTagSetCursor returns a instance of TagSetCursor.
func (*TagSetCursor) Fields ¶
func (tsc *TagSetCursor) Fields() map[string]interface{}
Fields returns the current fields of the current cursor
func (*TagSetCursor) Init ¶
func (tsc *TagSetCursor) Init(seek int64)
func (*TagSetCursor) Next ¶
func (tsc *TagSetCursor) Next(tmin, tmax int64) (int64, interface{})
Next returns the next matching series-key, timestamp byte slice and meta tags for the tagset. Filtering is enforced on the values. If there is no matching value, then a nil result is returned.
func (*TagSetCursor) Tags ¶
func (tsc *TagSetCursor) Tags() map[string]string
Tags returns the current tags of the current cursor if there is no current currsor, it returns nil
type TagSetCursors ¶
type TagSetCursors []*TagSetCursor
TagSetCursors represents a sortable slice of TagSetCursors.
func (TagSetCursors) Keys ¶
func (a TagSetCursors) Keys() []string
func (TagSetCursors) Len ¶
func (a TagSetCursors) Len() int
func (TagSetCursors) Less ¶
func (a TagSetCursors) Less(i, j int) bool
func (TagSetCursors) Swap ¶
func (a TagSetCursors) Swap(i, j int)
type TagsCursor ¶
type TagsCursor struct {
// contains filtered or unexported fields
}
TagsCursor is a cursor with attached tags and filter.
func NewTagsCursor ¶
NewTagsCursor returns a new instance of a series cursor.
func (*TagsCursor) Next ¶
func (c *TagsCursor) Next() (int64, interface{})
Next returns the next timestamp and value from the cursor.
func (*TagsCursor) SeekTo ¶
func (c *TagsCursor) SeekTo(seek int64) (int64, interface{})
Seek positions returning the key and value at that key.
type Tx ¶ added in v0.9.3
type Tx interface { io.WriterTo Size() int64 Commit() error Rollback() error Cursor(series string, fields []string, dec *FieldCodec, ascending bool) Cursor }
Tx represents a transaction.
type UnmarshalFunc ¶
UnmarshalFunc represents a function that can take bytes from a mapper from remote server and marshal it into an interface the reducer can use
func InitializeUnmarshaller ¶
func InitializeUnmarshaller(c *influxql.Call) (UnmarshalFunc, error)
Source Files ¶
Directories ¶
Path | Synopsis |
---|---|
wal
Package WAL implements a write ahead log optimized for write throughput that can be put in front of the database index.
|
Package WAL implements a write ahead log optimized for write throughput that can be put in front of the database index. |
Package internal is a generated protocol buffer package.
|
Package internal is a generated protocol buffer package. |