Documentation
¶
Overview ¶
Copyright 2024 Huawei Cloud Computing Technologies Co., Ltd.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
Copyright 2022 Huawei Cloud Computing Technologies Co., Ltd.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
Copyright 2023 Huawei Cloud Computing Technologies Co., Ltd.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
Index ¶
- Constants
- Variables
- func AddLocations(l *immutable.LocationCursor, files immutable.TableReaders, ...) error
- func AddLocationsWithFirstTime(l *immutable.LocationCursor, files immutable.TableReaders, ...) (int64, error)
- func AddLocationsWithInit(l *immutable.LocationCursor, files immutable.TableReaders, ...) error
- func AddLocationsWithLimit(l *immutable.LocationCursor, files immutable.TableReaders, ...) (int64, error)
- func AppendColumnTimes(bitmap []bool, column executor.Column, columnTimes []int64, ...)
- func AppendNilRowWithTime(rec *record.Record, t int64)
- func AppendRecWithNilRows(rec, re *record.Record, opt hybridqp.Options, ...)
- func CanNotAggOnSeriesFunc(m map[string]*influxql.Call) bool
- func FloatCountPromMerge(prevBuf, currBuf *floatColBuf)
- func FloatCountPromReduce(cv *record.ColVal, values []float64, start, end int) (int, float64, bool)
- func GenPeerPtFilePath(sh Shard, peersPtIDMap map[uint32]*NodeInfo, nodePath, fullPath string) []string
- func GenShardDirPath(metaClient meta.MetaClient, dbName string, ptId uint32) (map[uint32]*NodeInfo, error)
- func GetCtx(querySchema *executor.QuerySchema) (*idKeyCursorContext, error)
- func GetIntersectTimeRange(queryStartTime, queryEndTime, shardStartTime, shardEndTime int64) util.TimeRange
- func GetMaxTime(maxTime int64, rec *record.Record, isAscending bool) int64
- func GetMemUsageLimit() int32
- func GetMinTime(minTime int64, rec *record.Record, isAscending bool) int64
- func IsMemUsageExceeded() bool
- func MaxPromMerge(prevBuf, currBuf *floatColBuf)
- func MaxPromReduce(_ *record.ColVal, values []float64, start, end int) (int, float64, bool)
- func MinPromMerge(prevBuf, currBuf *floatColBuf)
- func MinPromReduce(_ *record.ColVal, values []float64, start, end int) (int, float64, bool)
- func NewAggregateCursor(input comm.KeyCursor, schema *executor.QuerySchema, ...) *aggregateCursor
- func NewAttachedIndexReader(ctx *indexContext, info *executor.AttachedIndexInfo, ...) *attachedIndexReader
- func NewChunkReader(outputRowDataType hybridqp.RowDataType, ops []hybridqp.ExprOptions, ...) executor.Processor
- func NewDetachedIndexReader(ctx *indexContext, obsOption *obs.ObsOptions, ...) *detachedIndexReader
- func NewDetachedLazyLoadIndexReader(ctx *indexContext, obsOption *obs.ObsOptions, ...) *detachedLazyLoadIndexReader
- func NewEngine(dataPath, walPath string, options netstorage.EngineOptions, ctx *meta.LoadCtx) (netstorage.Engine, error)
- func NewFencer(dataPath, walPath, db string, pt uint32) fencer
- func NewFileLoopCursor(ctx *idKeyCursorContext, span *tracing.Span, schema *executor.QuerySchema, ...) *fileLoopCursor
- func NewFileSequenceAggregator(schema hybridqp.Catalog, addPrefix bool, shardStartTime, shardEndTime int64) executor.Processor
- func NewFloatColFloatHeapReducer(inOrdinal, outOrdinal int, auxProcessors []*auxProcessor, ...) *floatColFloatHeapReducer
- func NewIndexContext(readBatch bool, batchCount int, schema hybridqp.Catalog, shardPath string) *indexContext
- func NewIntegerColIntegerHeapReducer(inOrdinal, outOrdinal int, auxProcessors []*auxProcessor, ...) *integerColIntegerHeapReducer
- func NewLimitCursor(schema *executor.QuerySchema, ...) *limitCursor
- func NewRecordSchema(querySchema *executor.QuerySchema, auxTags []string, schema record.Schemas, ...) ([]string, record.Schemas)
- func NewSeriesInfoPool(num int64) *filesInfoPool
- func NewShard(dataPath, walPath string, lockPath *string, ident *meta.ShardIdentifier, ...) *shard
- func NewTagSetCursorForTest(schema *executor.QuerySchema, seriesN int) *tagSetCursor
- func NewTopNLinkedList(n int, ascending bool) *topNLinkedList
- func NewTsspSequenceReader(outputRowDataType hybridqp.RowDataType, ops []hybridqp.ExprOptions, ...) executor.Processor
- func NewWriteIntoStorageTransform(outputRowDataType hybridqp.RowDataType, ops []hybridqp.ExprOptions, ...) executor.Processor
- func RecordCutNormal(start, end int, src, dst *record.Record)
- func RegistryAggOp(name string, aggOp AggOperator)
- func RegistryPromFunction(name string, aggOp PromFunction)
- func SetFullCompColdDuration(d time.Duration)
- func SetNextMethod(cursor comm.KeyCursor)
- type AggOperator
- type AggParams
- type AggTagSetCursor
- func (s *AggTagSetCursor) Close() error
- func (s *AggTagSetCursor) EndSpan()
- func (s *AggTagSetCursor) GetIndex(t int64) int64
- func (s *AggTagSetCursor) GetSchema() record.Schemas
- func (s *AggTagSetCursor) Init()
- func (s *AggTagSetCursor) Name() string
- func (s *AggTagSetCursor) Next() (*record.Record, comm.SeriesInfoIntf, error)
- func (s *AggTagSetCursor) NextAggData() (*record.Record, *comm.FileInfo, error)
- func (s *AggTagSetCursor) NextWithMultipleSeries() (*record.Record, comm.SeriesInfoIntf, error)
- func (s *AggTagSetCursor) NextWithSingleSeries() (*record.Record, comm.SeriesInfoIntf, error)
- func (s *AggTagSetCursor) RecordInit() error
- func (s *AggTagSetCursor) SetOps(ops []*comm.CallOption)
- func (s *AggTagSetCursor) SetParaForTest(schema record.Schemas)
- func (s *AggTagSetCursor) SetSchema(schema record.Schemas)
- func (s *AggTagSetCursor) SinkPlan(plan hybridqp.QueryNode)
- func (s *AggTagSetCursor) StartSpan(span *tracing.Span)
- func (s *AggTagSetCursor) TagAuxHandler(re *record.Record, start, end int)
- func (s *AggTagSetCursor) TimeWindowsInit()
- func (s *AggTagSetCursor) UpdateRec(recRow, chunkRow int)
- type Backup
- func (s *Backup) BackupPt(dbName string, ptId uint32) error
- func (s *Backup) FullBackup(sh Shard, dataPath, nodePath string, peersPtIDMap map[uint32]*NodeInfo) error
- func (s *Backup) FullBackupTableFile(sh Shard, t immutable.TablesStore, peersPtIDMap map[uint32]*NodeInfo, ...) ([][]string, error)
- func (s *Backup) IncBackup(sh Shard, dataPath, nodePath string, peersPtIDMap map[uint32]*NodeInfo) error
- func (s *Backup) IncBackupTableFile(sh Shard, t immutable.TablesStore, peersPtIDMap map[uint32]*NodeInfo, ...) ([][]string, [][]string, error)
- func (s *Backup) RunBackupData() error
- type BasePromOp
- type ChunkMetaByField
- type ChunkMetaByFieldIters
- type ChunkReader
- func (r *ChunkReader) Abort()
- func (r *ChunkReader) Close()
- func (r *ChunkReader) Create(plan executor.LogicalPlan, opt *query.ProcessorOptions) (executor.Processor, error)
- func (r *ChunkReader) Explain() []executor.ValuePair
- func (r *ChunkReader) GetInputNumber(executor.Port) int
- func (r *ChunkReader) GetInputs() executor.Ports
- func (r *ChunkReader) GetOutputNumber(executor.Port) int
- func (r *ChunkReader) GetOutputs() executor.Ports
- func (r *ChunkReader) IsSink() bool
- func (r *ChunkReader) Name() string
- func (r *ChunkReader) Release() error
- func (r *ChunkReader) Work(ctx context.Context) error
- type CoProcessor
- type CoProcessorImpl
- type ColumnStoreReader
- func (r *ColumnStoreReader) Abort()
- func (r *ColumnStoreReader) Close()
- func (r *ColumnStoreReader) Explain() []executor.ValuePair
- func (r *ColumnStoreReader) FragmentCount() int
- func (r *ColumnStoreReader) GetInputNumber(_ executor.Port) int
- func (r *ColumnStoreReader) GetInputs() executor.Ports
- func (r *ColumnStoreReader) GetOutputNumber(_ executor.Port) int
- func (r *ColumnStoreReader) GetOutputs() executor.Ports
- func (r *ColumnStoreReader) IsSink() bool
- func (r *ColumnStoreReader) Name() string
- func (r *ColumnStoreReader) Release() error
- func (r *ColumnStoreReader) Run(ctx context.Context) (iterCount, rowCountAfterFilter int, err error)
- func (r *ColumnStoreReader) Work(ctx context.Context) error
- type ColumnStoreReaderCreator
- type Compactor
- func (c *Compactor) RegisterShard(sh *shard)
- func (c *Compactor) SetAllOutOfOrderMergeSwitch(en bool)
- func (c *Compactor) SetAllShardsCompactionSwitch(en bool)
- func (c *Compactor) SetSnapshotColdDuration(d time.Duration)
- func (c *Compactor) ShardCompactionSwitch(shid uint64, en bool)
- func (c *Compactor) ShardOutOfOrderMergeSwitch(shid uint64, en bool)
- func (c *Compactor) UnregisterShard(shardId uint64)
- type DBPTInfo
- func (dbPT *DBPTInfo) AddShard(id uint64, sh Shard)
- func (dbPT *DBPTInfo) NewShard(rp string, shardID uint64, timeRangeInfo *meta.ShardTimeRangeInfo, ...) (Shard, error)
- func (dbPT *DBPTInfo) OpenIndexes(opId uint64, rp string, engineType config.EngineType) error
- func (dbPT *DBPTInfo) OpenShards(opId uint64, rp string, durationInfos map[uint64]*meta.ShardDurationInfo, ...) error
- func (dbPT *DBPTInfo) SetOption(opt netstorage.EngineOptions)
- func (dbPT *DBPTInfo) SetParams(preload bool, lockPath *string, enableTagArray bool)
- func (dbPT *DBPTInfo) Shard(id uint64) Shard
- func (dbPT *DBPTInfo) ShardIds(tr *influxql.TimeRange) []uint64
- func (dbPT *DBPTInfo) ShardNoLock(id uint64) Shard
- type DataBlockInfo
- type DetachedMetaInfo
- type DownSampleFilesInfo
- type EndPointPair
- type Engine
- func (e *Engine) Assign(opId uint64, nodeId uint64, db string, ptId uint32, ver uint64, ...) error
- func (e *Engine) ChangeShardTierToWarm(db string, ptId uint32, shardID uint64) error
- func (e *Engine) CheckPtsRemovedDone() bool
- func (e *Engine) ClearIndexCache(db string, ptId uint32, indexID uint64) error
- func (e *Engine) Close() error
- func (e *Engine) CreateDBPT(db string, pt uint32, enableTagArray bool)
- func (e *Engine) CreateLogicalPlan(ctx context.Context, db string, ptId uint32, shardID uint64, ...) (hybridqp.QueryNode, error)
- func (e *Engine) CreateShard(db, rp string, ptId uint32, shardID uint64, ...) error
- func (e *Engine) CreateShowTagValuesPlan(db string, ptIDs []uint32, tr *influxql.TimeRange) netstorage.ShowTagValuesPlan
- func (e *Engine) Databases() []string
- func (e *Engine) DbPTRef(db string, ptId uint32) error
- func (e *Engine) DbPTUnref(db string, ptId uint32)
- func (e *Engine) DeleteDatabase(db string, ptId uint32) (err error)
- func (e *Engine) DeleteIndex(db string, ptId uint32, indexID uint64) error
- func (e *Engine) DeleteShard(db string, ptId uint32, shardID uint64) error
- func (e *Engine) DropMeasurement(db string, rp string, name string, shardIds []uint64) error
- func (e *Engine) DropRetentionPolicy(db string, rp string, ptId uint32) error
- func (e *Engine) DropSeries(database string, sources []influxql.Source, ptId []uint32, ...) (int, error)
- func (e *Engine) ExpiredCacheIndexes() []*meta2.IndexIdentifier
- func (e *Engine) ExpiredIndexes() []*meta2.IndexIdentifier
- func (e *Engine) ExpiredShards() []*meta2.ShardIdentifier
- func (e *Engine) FetchShardsNeedChangeStore() (shardsToWarm, shardsToCold []*meta2.ShardIdentifier)
- func (e *Engine) ForceFlush()
- func (e *Engine) GetDBPtIds() map[string][]uint32
- func (e *Engine) GetDownSamplePolicy(key string) *meta2.StoreDownSamplePolicy
- func (e *Engine) GetIndexInfo(db string, ptId uint32, shardID uint64, schema *executor.QuerySchema) (*executor.AttachedIndexInfo, error)
- func (e *Engine) GetLockFile() string
- func (e *Engine) GetShard(db string, ptId uint32, shardID uint64) (Shard, error)
- func (e *Engine) GetShardDownSampleLevel(db string, ptId uint32, shardID uint64) int
- func (e *Engine) GetShardDownSamplePolicyInfos(meta interface{ ... }) ([]*meta2.ShardDownSamplePolicyInfo, error)
- func (e *Engine) GetShardSplitPoints(db string, ptId uint32, shardID uint64, idxes []int64) ([]string, error)
- func (e *Engine) HierarchicalStorage(db string, ptId uint32, shardID uint64) bool
- func (s *Engine) InitLogStoreCtx(querySchema *executor.QuerySchema) (*idKeyCursorContext, error)
- func (e *Engine) LogicalPlanCost(db string, ptId uint32, sources influxql.Sources, opt query.ProcessorOptions) (hybridqp.LogicalPlanCost, error)
- func (e *Engine) Offload(opId uint64, db string, ptId uint32) error
- func (e *Engine) Open(durationInfos map[uint64]*meta2.ShardDurationInfo, ...) error
- func (e *Engine) PreAssign(opId uint64, db string, ptId uint32, ...) error
- func (e *Engine) PreOffload(opId uint64, db string, ptId uint32) error
- func (e *Engine) RollbackPreOffload(opId uint64, db string, ptId uint32) error
- func (e *Engine) RowCount(db string, ptId uint32, shardIDs []uint64, schema *executor.QuerySchema) (int64, error)
- func (e *Engine) ScanWithSparseIndex(ctx context.Context, db string, ptId uint32, shardIDs []uint64, ...) (executor.ShardsFragments, error)
- func (e *Engine) SendRaftMessage(database string, ptId uint64, msgs raftpb.Message) error
- func (e *Engine) SeriesCardinality(db string, ptIDs []uint32, namesWithVer [][]byte, condition influxql.Expr, ...) ([]meta2.MeasurementCardinalityInfo, error)
- func (e *Engine) SeriesExactCardinality(db string, ptIDs []uint32, measurements [][]byte, condition influxql.Expr, ...) (map[string]uint64, error)
- func (e *Engine) SeriesKeys(db string, ptIDs []uint32, measurements [][]byte, condition influxql.Expr, ...) ([]string, error)
- func (e *Engine) SetMetaClient(m meta.MetaClient)
- func (e *Engine) StartDownSampleTask(sdsp *meta2.ShardDownSamplePolicyInfo, schema []hybridqp.Catalog, ...) error
- func (e *Engine) Statistics(buffer []byte) ([]byte, error)
- func (e *Engine) StatisticsOps() []opsStat.OpsStatistic
- func (e *Engine) SysCtrl(req *netstorage.SysCtrlRequest) (map[string]string, error)
- func (e *Engine) TagKeys(db string, ptIDs []uint32, measurements [][]byte, condition influxql.Expr, ...) ([]string, error)
- func (e *Engine) TagValues(db string, ptIDs []uint32, tagKeys map[string][][]byte, ...) (netstorage.TablesTagSets, error)
- func (e *Engine) TagValuesCardinality(db string, ptIDs []uint32, tagKeys map[string][][]byte, ...) (map[string]uint64, error)
- func (e *Engine) UpdateDownSampleInfo(policies *meta2.DownSamplePoliciesInfoWithDbRp)
- func (e *Engine) UpdateShardDownSampleInfo(infos *meta2.ShardDownSampleUpdateInfos)
- func (e *Engine) UpdateShardDurationInfo(info *meta2.ShardDurationInfo) error
- func (e *Engine) UpdateStoreDownSamplePolicies(info *meta2.DownSamplePolicyInfo, ident string)
- func (e *Engine) WriteRec(db, mst string, ptId uint32, shardID uint64, rec *record.Record, ...) error
- func (e *Engine) WriteRows(db, rp string, ptId uint32, shardID uint64, rows []influx.Row, ...) error
- func (e *Engine) WriteToRaft(db, rp string, ptId uint32, tail []byte) error
- type FieldIter
- type FileSequenceAggregator
- func (r *FileSequenceAggregator) Aggregate()
- func (r *FileSequenceAggregator) AggregateSameSchema() error
- func (r *FileSequenceAggregator) Close()
- func (r *FileSequenceAggregator) Create(plan executor.LogicalPlan, _ *query.ProcessorOptions) (executor.Processor, error)
- func (r *FileSequenceAggregator) Explain() []executor.ValuePair
- func (r *FileSequenceAggregator) GetInputNumber(executor.Port) int
- func (r *FileSequenceAggregator) GetInputs() executor.Ports
- func (r *FileSequenceAggregator) GetOutputNumber(executor.Port) int
- func (r *FileSequenceAggregator) GetOutputs() executor.Ports
- func (r *FileSequenceAggregator) GetProcessors()
- func (r *FileSequenceAggregator) IsSink() bool
- func (r *FileSequenceAggregator) Name() string
- func (r *FileSequenceAggregator) Release() error
- func (r *FileSequenceAggregator) SendRecord(re *record.Record)
- func (r *FileSequenceAggregator) Work(ctx context.Context) error
- type FloatCountPromOp
- type FloatHeapItem
- type FloatIncAggMergeFunc
- type FloatIncAggReduceFunc
- type FloatPointItem
- type FloatRateMergeFunc
- type FloatRateReduceFunc
- type FloatRateUpdateFunc
- type FloatSliceMergeFunc
- type FloatSliceReduceFunc
- type HybridStoreReader
- func (r *HybridStoreReader) Abort()
- func (r *HybridStoreReader) Close()
- func (r *HybridStoreReader) CreateCursors() error
- func (r *HybridStoreReader) Explain() []executor.ValuePair
- func (r *HybridStoreReader) GetInputNumber(_ executor.Port) int
- func (r *HybridStoreReader) GetInputs() executor.Ports
- func (r *HybridStoreReader) GetOutputNumber(_ executor.Port) int
- func (r *HybridStoreReader) GetOutputs() executor.Ports
- func (r *HybridStoreReader) IsSink() bool
- func (r *HybridStoreReader) Name() string
- func (r *HybridStoreReader) Release() error
- func (r *HybridStoreReader) Work(ctx context.Context) (err error)
- type IndexReader
- type InstantVectorCursor
- func (c *InstantVectorCursor) Close() error
- func (c *InstantVectorCursor) EndSpan()
- func (c *InstantVectorCursor) GetSchema() record.Schemas
- func (c *InstantVectorCursor) Name() string
- func (c *InstantVectorCursor) Next() (*record.Record, comm.SeriesInfoIntf, error)
- func (c *InstantVectorCursor) NextAggData() (*record.Record, *comm.FileInfo, error)
- func (c *InstantVectorCursor) SetOps(ops []*comm.CallOption)
- func (c *InstantVectorCursor) SetSchema(inSchema, outSchema record.Schemas, exprOpt []hybridqp.ExprOptions)
- func (c *InstantVectorCursor) SinkPlan(plan hybridqp.QueryNode)
- func (c *InstantVectorCursor) StartSpan(span *tracing.Span)
- type IntegerHeapItem
- type IntegerPointItem
- type LogReplay
- type LogReplays
- type LogWriter
- type LogWriters
- type MaxPromOp
- type MetaIndexIterator
- type MinPromOp
- type NodeInfo
- type PreAggTagSetCursor
- func (s *PreAggTagSetCursor) Close() error
- func (s *PreAggTagSetCursor) EndSpan()
- func (s *PreAggTagSetCursor) GetSchema() record.Schemas
- func (s *PreAggTagSetCursor) Name() string
- func (s *PreAggTagSetCursor) Next() (*record.Record, comm.SeriesInfoIntf, error)
- func (s *PreAggTagSetCursor) NextAggData() (*record.Record, *comm.FileInfo, error)
- func (s *PreAggTagSetCursor) RecordInitPreAgg() error
- func (s *PreAggTagSetCursor) SetOps(ops []*comm.CallOption)
- func (s *PreAggTagSetCursor) SetSchema(schema record.Schemas)
- func (s *PreAggTagSetCursor) SinkPlan(plan hybridqp.QueryNode)
- func (s *PreAggTagSetCursor) StartSpan(span *tracing.Span)
- type PredictLinearOp
- type PromFuncParam
- type PromFunction
- type PromFunctionFactory
- type PtNNLock
- type QuantileOverTime
- type RangeVectorCursor
- func (c *RangeVectorCursor) Close() error
- func (c *RangeVectorCursor) EndSpan()
- func (c *RangeVectorCursor) GetSchema() record.Schemas
- func (c *RangeVectorCursor) Name() string
- func (c *RangeVectorCursor) Next() (*record.Record, comm.SeriesInfoIntf, error)
- func (c *RangeVectorCursor) NextAggData() (*record.Record, *comm.FileInfo, error)
- func (c *RangeVectorCursor) SetOps(ops []*comm.CallOption)
- func (c *RangeVectorCursor) SetSchema(inSchema, outSchema record.Schemas, exprOpt []hybridqp.ExprOptions)
- func (c *RangeVectorCursor) SinkPlan(plan hybridqp.QueryNode)
- func (c *RangeVectorCursor) StartSpan(span *tracing.Span)
- type Reducer
- type ReducerEndpoint
- type ReducerParams
- type Routine
- type RoutineImpl
- type SeriesIter
- type Shard
- type ShardMoveFileInfo
- type ShardStatus
- type ShowTagValuesPlan
- type StdDevOverTime
- type StdVarOverTime
- type Storage
- type StreamDetachedReader
- func (t *StreamDetachedReader) Close() error
- func (t *StreamDetachedReader) EndSpan()
- func (t *StreamDetachedReader) GetSchema() record.Schemas
- func (r *StreamDetachedReader) Init() (err error)
- func (t *StreamDetachedReader) Name() string
- func (t *StreamDetachedReader) Next() (*record.Record, comm.SeriesInfoIntf, error)
- func (t *StreamDetachedReader) NextAggData() (*record.Record, *comm.FileInfo, error)
- func (t *StreamDetachedReader) SetOps(ops []*comm.CallOption)
- func (t *StreamDetachedReader) SinkPlan(plan hybridqp.QueryNode)
- func (t *StreamDetachedReader) StartSpan(span *tracing.Span)
- func (t *StreamDetachedReader) UpdateTime(time int64)
- type TagSetCursorItem
- type TierInfo
- type TsspSequenceReader
- func (r *TsspSequenceReader) Close()
- func (r *TsspSequenceReader) Create(plan executor.LogicalPlan, opt *query.ProcessorOptions) (executor.Processor, error)
- func (r *TsspSequenceReader) Explain() []executor.ValuePair
- func (r *TsspSequenceReader) GetInputNumber(executor.Port) int
- func (r *TsspSequenceReader) GetInputs() executor.Ports
- func (r *TsspSequenceReader) GetOutputNumber(executor.Port) int
- func (r *TsspSequenceReader) GetOutputs() executor.Ports
- func (r *TsspSequenceReader) IsSink() bool
- func (r *TsspSequenceReader) Name() string
- func (r *TsspSequenceReader) Release() error
- func (r *TsspSequenceReader) Work(ctx context.Context) error
- type WAL
- type WalRecordType
- type WriteIntoStorageTransform
- func (r *WriteIntoStorageTransform) Close()
- func (r *WriteIntoStorageTransform) Create(plan executor.LogicalPlan, opt *query.ProcessorOptions) (executor.Processor, error)
- func (r *WriteIntoStorageTransform) EndFile() error
- func (r *WriteIntoStorageTransform) Explain() []executor.ValuePair
- func (r *WriteIntoStorageTransform) GetClosed() chan struct{}
- func (r *WriteIntoStorageTransform) GetInputNumber(executor.Port) int
- func (r *WriteIntoStorageTransform) GetInputs() executor.Ports
- func (r *WriteIntoStorageTransform) GetOutputNumber(executor.Port) int
- func (r *WriteIntoStorageTransform) GetOutputs() executor.Ports
- func (r *WriteIntoStorageTransform) GetRowCount() int
- func (r *WriteIntoStorageTransform) InitFile(sRecord *executor.SeriesRecord) error
- func (r *WriteIntoStorageTransform) Name() string
- func (r *WriteIntoStorageTransform) Release() error
- func (r *WriteIntoStorageTransform) SetTaskId(taskID int)
- func (r *WriteIntoStorageTransform) Work(ctx context.Context) error
Constants ¶
const ( MetaIndexAndBlockIdDuration = "meta_index_duration" PrimaryKeyDuration = "primary_key_duration" )
const ( HybridStoreReaderChunkNum = 7 SegmentBatchCount = 128 )
const ( SidSequenceReaderRecordNum = 6 SequenceAggRecordNum = 3 )
const ( MaxRetryUpdateOnShardNum = 4 CRCLen = 4 BufferSize = 1024 * 1024 // OBSFileExtension is the extension used for OBS files. OBSFileExtension = ".init" )
const ( Failpoint = "failpoint" BackgroundReadLimiter = "backgroundReadLimiter" )
const ( TierLeveMem = 1 // in memory TierLeveLocalDisk = 2 TierLeveObjectStorage = 3 )
const ( DefaultFileSize = 10 * 1024 * 1024 WALFileSuffixes = "wal" WalRecordHeadSize = 1 + 4 WalCompBufSize = 256 * 1024 WalCompMaxBufSize = 2 * 1024 * 1024 )
const ( WriteWalUnKnownType = iota WriteWalLineProtocol WriteWalArrowFlight WriteWalEnd )
const ColumnStoreReaderChunkNum = 7
const ColumnStoreReaderRecordNum = 7
const DefaultUploadFrequence = 500 * time.Millisecond
const (
IncDataSegmentNum = 16
)
const MaxFileInfoSize = 1024
Variables ¶
var ( FileLoopCursorPool = record.NewRecordPool(record.FileLoopCursorPool) FileCursorPool = record.NewRecordPool(record.FileCursorPool) FileCursorValidRowPool = record.NewRecordPool(record.FileCursorValidRowPool) FileCursorFilterRecordPool = record.NewRecordPool(record.FileCursorFilterRecordPool) )
var ( AggPool = record.NewRecordPool(record.AggPool) SeriesPool = record.NewRecordPool(record.SeriesPool) SeriesLoopPool = record.NewRecordPool(record.SeriesLoopPool) TsmMergePool = record.NewRecordPool(record.TsmMergePool) )
var ( TsspSequencePool = record.NewRecordPool(record.TsspSequencePool) SequenceAggPool = record.NewRecordPool(record.SequenceAggPool) )
var AggFactory = make(map[string]AggOperator)
var AppendManyNils map[int]func(colVal *record.ColVal, count int)
var DeleteDatabaseTimeout = time.Second * 15
var (
DownSampleWriteDrop = true
)
var (
IntervalRecordPool = record.NewRecordPool(record.IntervalRecordPool)
)
var (
RecordIteratorPool = &sync.Pool{}
)
Functions ¶
func AddLocations ¶ added in v1.3.0
func AddLocations(l *immutable.LocationCursor, files immutable.TableReaders, ctx *idKeyCursorContext, sid uint64, metaCtx *immutable.ChunkMetaContext) error
func AddLocationsWithFirstTime ¶
func AddLocationsWithFirstTime(l *immutable.LocationCursor, files immutable.TableReaders, ctx *idKeyCursorContext, sid uint64) (int64, error)
func AddLocationsWithInit ¶
func AddLocationsWithInit(l *immutable.LocationCursor, files immutable.TableReaders, ctx *idKeyCursorContext, sid uint64) error
func AddLocationsWithLimit ¶
func AddLocationsWithLimit(l *immutable.LocationCursor, files immutable.TableReaders, ctx *idKeyCursorContext, sid uint64) (int64, error)
func AppendColumnTimes ¶
func AppendNilRowWithTime ¶ added in v1.0.0
func AppendRecWithNilRows ¶ added in v1.0.0
func FloatCountPromMerge ¶ added in v1.3.0
func FloatCountPromMerge(prevBuf, currBuf *floatColBuf)
func FloatCountPromReduce ¶ added in v1.3.0
func GenPeerPtFilePath ¶ added in v1.3.0
func GenShardDirPath ¶ added in v1.3.0
func GetCtx ¶ added in v1.3.0
func GetCtx(querySchema *executor.QuerySchema) (*idKeyCursorContext, error)
func GetIntersectTimeRange ¶ added in v1.3.0
func GetIntersectTimeRange(queryStartTime, queryEndTime, shardStartTime, shardEndTime int64) util.TimeRange
GetIntersectTimeRange used to get intersection of the query time and shard time
func GetMemUsageLimit ¶ added in v1.1.0
func GetMemUsageLimit() int32
func IsMemUsageExceeded ¶ added in v1.1.0
func IsMemUsageExceeded() bool
func MaxPromMerge ¶ added in v1.3.0
func MaxPromMerge(prevBuf, currBuf *floatColBuf)
func MaxPromReduce ¶ added in v1.3.0
func MinPromMerge ¶ added in v1.3.0
func MinPromMerge(prevBuf, currBuf *floatColBuf)
func MinPromReduce ¶ added in v1.3.0
func NewAggregateCursor ¶
func NewAggregateCursor(input comm.KeyCursor, schema *executor.QuerySchema, globalPool *record.RecordPool, hasAuxTags bool) *aggregateCursor
func NewAttachedIndexReader ¶ added in v1.2.0
func NewAttachedIndexReader(ctx *indexContext, info *executor.AttachedIndexInfo, readerCtx *immutable.FileReaderContext) *attachedIndexReader
func NewChunkReader ¶
func NewChunkReader(outputRowDataType hybridqp.RowDataType, ops []hybridqp.ExprOptions, seriesPlan hybridqp.QueryNode, schema *executor.QuerySchema, cursors []interface{}, state bool) executor.Processor
func NewDetachedIndexReader ¶ added in v1.2.0
func NewDetachedIndexReader(ctx *indexContext, obsOption *obs.ObsOptions, readerCtx *immutable.FileReaderContext) *detachedIndexReader
func NewDetachedLazyLoadIndexReader ¶ added in v1.3.0
func NewDetachedLazyLoadIndexReader(ctx *indexContext, obsOption *obs.ObsOptions, readerCtx *immutable.FileReaderContext) *detachedLazyLoadIndexReader
func NewEngine ¶
func NewEngine(dataPath, walPath string, options netstorage.EngineOptions, ctx *meta.LoadCtx) (netstorage.Engine, error)
func NewFileLoopCursor ¶
func NewFileLoopCursor(ctx *idKeyCursorContext, span *tracing.Span, schema *executor.QuerySchema, tagSet *tsi.TagSetInfo, start, step int, s *shard) *fileLoopCursor
func NewFileSequenceAggregator ¶ added in v1.0.0
func NewFloatColFloatHeapReducer ¶
func NewFloatColFloatHeapReducer(inOrdinal, outOrdinal int, auxProcessors []*auxProcessor, floatHeapItem *FloatHeapItem) *floatColFloatHeapReducer
func NewIndexContext ¶ added in v1.2.0
func NewIntegerColIntegerHeapReducer ¶
func NewIntegerColIntegerHeapReducer(inOrdinal, outOrdinal int, auxProcessors []*auxProcessor, integerHeapItem *IntegerHeapItem) *integerColIntegerHeapReducer
func NewLimitCursor ¶
func NewLimitCursor(schema *executor.QuerySchema, helper func(start, end int, src, des *record.Record)) *limitCursor
func NewRecordSchema ¶
func NewSeriesInfoPool ¶
func NewSeriesInfoPool(num int64) *filesInfoPool
func NewShard ¶
func NewShard(dataPath, walPath string, lockPath *string, ident *meta.ShardIdentifier, durationInfo *meta.DurationDescriptor, tr *meta.TimeRangeInfo, options netstorage.EngineOptions, engineType config.EngineType, ch chan []immutable.FileInfoExtend) *shard
func NewTagSetCursorForTest ¶
func NewTagSetCursorForTest(schema *executor.QuerySchema, seriesN int) *tagSetCursor
NewTagSetCursorForTest for ut test, will remove later
func NewTopNLinkedList ¶
func NewTsspSequenceReader ¶ added in v1.0.0
func NewTsspSequenceReader(outputRowDataType hybridqp.RowDataType, ops []hybridqp.ExprOptions, seriesPlan hybridqp.QueryNode, source influxql.Sources, schema *executor.QuerySchema, files *immutable.TSSPFiles, newSeqs []uint64, stop chan struct{}) executor.Processor
func NewWriteIntoStorageTransform ¶ added in v1.0.0
func RecordCutNormal ¶
func RegistryAggOp ¶ added in v1.3.0
func RegistryAggOp(name string, aggOp AggOperator)
func RegistryPromFunction ¶ added in v1.3.0
func RegistryPromFunction(name string, aggOp PromFunction)
func SetFullCompColdDuration ¶
func SetNextMethod ¶ added in v1.0.0
SetNextMethod for test
Types ¶
type AggOperator ¶ added in v1.3.0
func GetAggOperator ¶ added in v1.3.0
func GetAggOperator(name string) AggOperator
type AggParams ¶ added in v1.3.0
type AggParams struct {
// contains filtered or unexported fields
}
func NewAggParams ¶ added in v1.3.0
func NewAggParams(inSchema, outSchema record.Schemas, opt hybridqp.ExprOptions, auxOp []*auxProcessor) *AggParams
type AggTagSetCursor ¶
type AggTagSetCursor struct {
// contains filtered or unexported fields
}
func NewAggTagSetCursor ¶
func NewAggTagSetCursor(schema *executor.QuerySchema, ctx *idKeyCursorContext, itr comm.KeyCursor, singleSeries bool) *AggTagSetCursor
func (*AggTagSetCursor) Close ¶
func (s *AggTagSetCursor) Close() error
func (*AggTagSetCursor) EndSpan ¶
func (s *AggTagSetCursor) EndSpan()
func (*AggTagSetCursor) GetIndex ¶ added in v1.0.0
func (s *AggTagSetCursor) GetIndex(t int64) int64
func (*AggTagSetCursor) GetSchema ¶
func (s *AggTagSetCursor) GetSchema() record.Schemas
func (*AggTagSetCursor) Init ¶
func (s *AggTagSetCursor) Init()
func (*AggTagSetCursor) Name ¶
func (s *AggTagSetCursor) Name() string
func (*AggTagSetCursor) Next ¶
func (s *AggTagSetCursor) Next() (*record.Record, comm.SeriesInfoIntf, error)
func (*AggTagSetCursor) NextAggData ¶
func (*AggTagSetCursor) NextWithMultipleSeries ¶ added in v1.0.0
func (s *AggTagSetCursor) NextWithMultipleSeries() (*record.Record, comm.SeriesInfoIntf, error)
func (*AggTagSetCursor) NextWithSingleSeries ¶ added in v1.0.0
func (s *AggTagSetCursor) NextWithSingleSeries() (*record.Record, comm.SeriesInfoIntf, error)
func (*AggTagSetCursor) RecordInit ¶
func (s *AggTagSetCursor) RecordInit() error
func (*AggTagSetCursor) SetOps ¶
func (s *AggTagSetCursor) SetOps(ops []*comm.CallOption)
func (*AggTagSetCursor) SetParaForTest ¶ added in v1.0.0
func (s *AggTagSetCursor) SetParaForTest(schema record.Schemas)
func (*AggTagSetCursor) SetSchema ¶
func (s *AggTagSetCursor) SetSchema(schema record.Schemas)
func (*AggTagSetCursor) SinkPlan ¶
func (s *AggTagSetCursor) SinkPlan(plan hybridqp.QueryNode)
func (*AggTagSetCursor) StartSpan ¶
func (s *AggTagSetCursor) StartSpan(span *tracing.Span)
func (*AggTagSetCursor) TagAuxHandler ¶
func (s *AggTagSetCursor) TagAuxHandler(re *record.Record, start, end int)
func (*AggTagSetCursor) TimeWindowsInit ¶
func (s *AggTagSetCursor) TimeWindowsInit()
func (*AggTagSetCursor) UpdateRec ¶ added in v1.0.0
func (s *AggTagSetCursor) UpdateRec(recRow, chunkRow int)
type Backup ¶ added in v1.3.0
type Backup struct { Name string IsInc bool IsRemote bool OnlyBackupMater bool BackupPath string BackupLogInfo *backup.BackupLogInfo Engine *Engine IsAborted bool // contains filtered or unexported fields }
func (*Backup) FullBackup ¶ added in v1.3.0
func (*Backup) FullBackupTableFile ¶ added in v1.3.0
func (*Backup) IncBackupTableFile ¶ added in v1.3.0
func (*Backup) RunBackupData ¶ added in v1.3.0
type BasePromOp ¶ added in v1.3.0
type BasePromOp struct {
// contains filtered or unexported fields
}
func NewBasePromOp ¶ added in v1.3.0
func NewBasePromOp(name string, fn floatColFloatReduce, fv floatColFloatMerge) *BasePromOp
func (*BasePromOp) CreateRoutine ¶ added in v1.3.0
func (c *BasePromOp) CreateRoutine(p *AggParams) (Routine, error)
type ChunkMetaByField ¶ added in v1.0.0
type ChunkMetaByField struct {
// contains filtered or unexported fields
}
ChunkMetaByField build from each chunkmeta. ChunkMetaByField obtain data by sid column by column
func NewChunkMetaByField ¶ added in v1.0.0
func NewChunkMetaByField(file immutable.TSSPFile, fieldIter *FieldIter, chunkMeta immutable.ChunkMeta, recordPool *record.CircularRecordPool) *ChunkMetaByField
type ChunkMetaByFieldIters ¶ added in v1.0.0
type ChunkMetaByFieldIters struct {
// contains filtered or unexported fields
}
ChunkMetaByFieldIters is the iterator of ChunkMetaByField.
func NewChunkMetaByFieldIters ¶ added in v1.0.0
func NewChunkMetaByFieldIters(chunkMetas []immutable.ChunkMeta, file immutable.TSSPFile, fieldIter *FieldIter, recordPool *record.CircularRecordPool) *ChunkMetaByFieldIters
type ChunkReader ¶
type ChunkReader struct { executor.BaseProcessor Output *executor.ChunkPort ResultChunkPool *executor.CircularChunkPool // contains filtered or unexported fields }
func (*ChunkReader) Abort ¶ added in v1.0.1
func (r *ChunkReader) Abort()
func (*ChunkReader) Close ¶
func (r *ChunkReader) Close()
func (*ChunkReader) Create ¶
func (r *ChunkReader) Create(plan executor.LogicalPlan, opt *query.ProcessorOptions) (executor.Processor, error)
func (*ChunkReader) Explain ¶
func (r *ChunkReader) Explain() []executor.ValuePair
func (*ChunkReader) GetInputNumber ¶
func (r *ChunkReader) GetInputNumber(executor.Port) int
func (*ChunkReader) GetInputs ¶
func (r *ChunkReader) GetInputs() executor.Ports
func (*ChunkReader) GetOutputNumber ¶
func (r *ChunkReader) GetOutputNumber(executor.Port) int
func (*ChunkReader) GetOutputs ¶
func (r *ChunkReader) GetOutputs() executor.Ports
func (*ChunkReader) IsSink ¶
func (r *ChunkReader) IsSink() bool
func (*ChunkReader) Name ¶
func (r *ChunkReader) Name() string
func (*ChunkReader) Release ¶
func (r *ChunkReader) Release() error
type CoProcessor ¶
type CoProcessor interface {
WorkOnRecord(*record.Record, *record.Record, *ReducerParams)
}
type CoProcessorImpl ¶
type CoProcessorImpl struct {
Routines []Routine
}
func NewCoProcessorImpl ¶
func NewCoProcessorImpl(routines ...Routine) *CoProcessorImpl
func (*CoProcessorImpl) AppendRoutine ¶
func (p *CoProcessorImpl) AppendRoutine(routines ...Routine)
func (*CoProcessorImpl) WorkOnRecord ¶
func (p *CoProcessorImpl) WorkOnRecord(in *record.Record, out *record.Record, params *ReducerParams)
type ColumnStoreReader ¶ added in v1.1.0
type ColumnStoreReader struct { executor.BaseProcessor // contains filtered or unexported fields }
func NewColumnStoreReader ¶ added in v1.1.0
func NewColumnStoreReader(plan hybridqp.QueryNode, frags executor.ShardsFragments) *ColumnStoreReader
func (*ColumnStoreReader) Abort ¶ added in v1.1.0
func (r *ColumnStoreReader) Abort()
func (*ColumnStoreReader) Close ¶ added in v1.1.0
func (r *ColumnStoreReader) Close()
func (*ColumnStoreReader) Explain ¶ added in v1.1.0
func (r *ColumnStoreReader) Explain() []executor.ValuePair
func (*ColumnStoreReader) FragmentCount ¶ added in v1.1.0
func (r *ColumnStoreReader) FragmentCount() int
func (*ColumnStoreReader) GetInputNumber ¶ added in v1.1.0
func (r *ColumnStoreReader) GetInputNumber(_ executor.Port) int
func (*ColumnStoreReader) GetInputs ¶ added in v1.1.0
func (r *ColumnStoreReader) GetInputs() executor.Ports
func (*ColumnStoreReader) GetOutputNumber ¶ added in v1.1.0
func (r *ColumnStoreReader) GetOutputNumber(_ executor.Port) int
func (*ColumnStoreReader) GetOutputs ¶ added in v1.1.0
func (r *ColumnStoreReader) GetOutputs() executor.Ports
func (*ColumnStoreReader) IsSink ¶ added in v1.1.0
func (r *ColumnStoreReader) IsSink() bool
func (*ColumnStoreReader) Name ¶ added in v1.1.0
func (r *ColumnStoreReader) Name() string
func (*ColumnStoreReader) Release ¶ added in v1.1.0
func (r *ColumnStoreReader) Release() error
type ColumnStoreReaderCreator ¶ added in v1.1.0
type ColumnStoreReaderCreator struct { }
func (*ColumnStoreReaderCreator) Create ¶ added in v1.1.0
func (c *ColumnStoreReaderCreator) Create(plan executor.LogicalPlan, _ *query.ProcessorOptions) (executor.Processor, error)
func (*ColumnStoreReaderCreator) CreateReader ¶ added in v1.1.0
type Compactor ¶
type Compactor struct {
// contains filtered or unexported fields
}
func NewCompactor ¶
func NewCompactor() *Compactor
func (*Compactor) RegisterShard ¶
func (c *Compactor) RegisterShard(sh *shard)
func (*Compactor) SetAllOutOfOrderMergeSwitch ¶
func (*Compactor) SetAllShardsCompactionSwitch ¶
func (*Compactor) SetSnapshotColdDuration ¶
func (*Compactor) ShardCompactionSwitch ¶
func (*Compactor) ShardOutOfOrderMergeSwitch ¶
func (*Compactor) UnregisterShard ¶
type DBPTInfo ¶
func NewDBPTInfo ¶
func NewDBPTInfo(db string, id uint32, dataPath, walPath string, ctx *metaclient.LoadCtx, ch chan []immutable.FileInfoExtend) *DBPTInfo
func (*DBPTInfo) NewShard ¶
func (dbPT *DBPTInfo) NewShard(rp string, shardID uint64, timeRangeInfo *meta.ShardTimeRangeInfo, client metaclient.MetaClient, engineType config.EngineType) (Shard, error)
func (*DBPTInfo) OpenIndexes ¶
func (*DBPTInfo) OpenShards ¶
func (dbPT *DBPTInfo) OpenShards(opId uint64, rp string, durationInfos map[uint64]*meta.ShardDurationInfo, loadStat int, client metaclient.MetaClient) error
func (*DBPTInfo) SetOption ¶
func (dbPT *DBPTInfo) SetOption(opt netstorage.EngineOptions)
func (*DBPTInfo) ShardNoLock ¶ added in v1.3.0
type DataBlockInfo ¶
type DataBlockInfo struct {
// contains filtered or unexported fields
}
type DetachedMetaInfo ¶ added in v1.2.0
type DetachedMetaInfo struct {
// contains filtered or unexported fields
}
func NewDetachedMetaInfo ¶ added in v1.2.0
func NewDetachedMetaInfo() *DetachedMetaInfo
type DownSampleFilesInfo ¶ added in v1.0.0
type EndPointPair ¶
type Engine ¶
type Engine struct { DBPartitions map[string]map[uint32]*DBPTInfo DownSamplePolicies map[string]*meta2.StoreDownSamplePolicy // contains filtered or unexported fields }
func (*Engine) Assign ¶ added in v1.0.0
func (e *Engine) Assign(opId uint64, nodeId uint64, db string, ptId uint32, ver uint64, durationInfos map[uint64]*meta2.ShardDurationInfo, dbBriefInfo *meta2.DatabaseBriefInfo, client metaclient.MetaClient, storage netstorage.StorageService) error
func (*Engine) ChangeShardTierToWarm ¶
func (*Engine) CheckPtsRemovedDone ¶ added in v1.1.1
func (*Engine) ClearIndexCache ¶ added in v1.2.0
func (*Engine) CreateLogicalPlan ¶
func (*Engine) CreateShard ¶
func (e *Engine) CreateShard(db, rp string, ptId uint32, shardID uint64, timeRangeInfo *meta2.ShardTimeRangeInfo, mstInfo *meta2.MeasurementInfo) error
func (*Engine) CreateShowTagValuesPlan ¶ added in v1.3.0
func (e *Engine) CreateShowTagValuesPlan(db string, ptIDs []uint32, tr *influxql.TimeRange) netstorage.ShowTagValuesPlan
func (*Engine) DeleteIndex ¶
func (*Engine) DeleteShard ¶
todo:need confirm
func (*Engine) DropMeasurement ¶
func (*Engine) DropRetentionPolicy ¶
func (*Engine) DropSeries ¶
func (*Engine) ExpiredCacheIndexes ¶ added in v1.2.0
func (e *Engine) ExpiredCacheIndexes() []*meta2.IndexIdentifier
func (*Engine) ExpiredIndexes ¶
func (e *Engine) ExpiredIndexes() []*meta2.IndexIdentifier
func (*Engine) ExpiredShards ¶
func (e *Engine) ExpiredShards() []*meta2.ShardIdentifier
func (*Engine) FetchShardsNeedChangeStore ¶
func (e *Engine) FetchShardsNeedChangeStore() (shardsToWarm, shardsToCold []*meta2.ShardIdentifier)
func (*Engine) ForceFlush ¶
func (e *Engine) ForceFlush()
func (*Engine) GetDBPtIds ¶ added in v1.3.0
func (*Engine) GetDownSamplePolicy ¶ added in v1.0.0
func (e *Engine) GetDownSamplePolicy(key string) *meta2.StoreDownSamplePolicy
func (*Engine) GetIndexInfo ¶ added in v1.2.0
func (e *Engine) GetIndexInfo(db string, ptId uint32, shardID uint64, schema *executor.QuerySchema) (*executor.AttachedIndexInfo, error)
func (*Engine) GetLockFile ¶
func (*Engine) GetShardDownSampleLevel ¶ added in v1.0.0
func (*Engine) GetShardDownSamplePolicyInfos ¶ added in v1.0.0
func (e *Engine) GetShardDownSamplePolicyInfos(meta interface { UpdateShardDownSampleInfo(Ident *meta2.ShardIdentifier) error }) ([]*meta2.ShardDownSamplePolicyInfo, error)
func (*Engine) GetShardSplitPoints ¶
func (*Engine) HierarchicalStorage ¶ added in v1.2.0
func (*Engine) InitLogStoreCtx ¶ added in v1.2.0
func (s *Engine) InitLogStoreCtx(querySchema *executor.QuerySchema) (*idKeyCursorContext, error)
func (*Engine) LogicalPlanCost ¶
func (e *Engine) LogicalPlanCost(db string, ptId uint32, sources influxql.Sources, opt query.ProcessorOptions) (hybridqp.LogicalPlanCost, error)
func (*Engine) Open ¶
func (e *Engine) Open(durationInfos map[uint64]*meta2.ShardDurationInfo, dbBriefInfos map[string]*meta2.DatabaseBriefInfo, m meta.MetaClient) error
func (*Engine) PreAssign ¶ added in v1.0.0
func (e *Engine) PreAssign(opId uint64, db string, ptId uint32, durationInfos map[uint64]*meta2.ShardDurationInfo, dbBriefInfo *meta2.DatabaseBriefInfo, client metaclient.MetaClient) error
func (*Engine) PreOffload ¶ added in v1.0.0
func (*Engine) RollbackPreOffload ¶ added in v1.0.0
func (*Engine) ScanWithSparseIndex ¶ added in v1.1.0
func (*Engine) SendRaftMessage ¶ added in v1.3.0
func (*Engine) SeriesCardinality ¶
func (*Engine) SeriesExactCardinality ¶
func (*Engine) SeriesKeys ¶
func (*Engine) SetMetaClient ¶ added in v1.3.0
func (e *Engine) SetMetaClient(m meta.MetaClient)
func (*Engine) StartDownSampleTask ¶ added in v1.0.0
func (*Engine) StatisticsOps ¶ added in v1.1.0
func (e *Engine) StatisticsOps() []opsStat.OpsStatistic
func (*Engine) SysCtrl ¶
func (e *Engine) SysCtrl(req *netstorage.SysCtrlRequest) (map[string]string, error)
func (*Engine) TagValuesCardinality ¶
func (*Engine) UpdateDownSampleInfo ¶ added in v1.0.0
func (e *Engine) UpdateDownSampleInfo(policies *meta2.DownSamplePoliciesInfoWithDbRp)
func (*Engine) UpdateShardDownSampleInfo ¶ added in v1.0.0
func (e *Engine) UpdateShardDownSampleInfo(infos *meta2.ShardDownSampleUpdateInfos)
func (*Engine) UpdateShardDurationInfo ¶
func (e *Engine) UpdateShardDurationInfo(info *meta2.ShardDurationInfo) error
func (*Engine) UpdateStoreDownSamplePolicies ¶ added in v1.0.0
func (e *Engine) UpdateStoreDownSamplePolicies(info *meta2.DownSamplePolicyInfo, ident string)
type FieldIter ¶ added in v1.0.0
type FieldIter struct {
// contains filtered or unexported fields
}
func NewFieldIter ¶ added in v1.0.0
func NewFieldIter(querySchema *executor.QuerySchema) *FieldIter
func (*FieldIter) GetRecordSchemas ¶ added in v1.0.0
type FileSequenceAggregator ¶ added in v1.0.0
type FileSequenceAggregator struct { executor.BaseProcessor Input *executor.SeriesRecordPort Output *executor.SeriesRecordPort // contains filtered or unexported fields }
func (*FileSequenceAggregator) Aggregate ¶ added in v1.0.0
func (r *FileSequenceAggregator) Aggregate()
func (*FileSequenceAggregator) AggregateSameSchema ¶ added in v1.0.0
func (r *FileSequenceAggregator) AggregateSameSchema() error
func (*FileSequenceAggregator) Close ¶ added in v1.0.0
func (r *FileSequenceAggregator) Close()
func (*FileSequenceAggregator) Create ¶ added in v1.0.0
func (r *FileSequenceAggregator) Create(plan executor.LogicalPlan, _ *query.ProcessorOptions) (executor.Processor, error)
func (*FileSequenceAggregator) Explain ¶ added in v1.0.0
func (r *FileSequenceAggregator) Explain() []executor.ValuePair
func (*FileSequenceAggregator) GetInputNumber ¶ added in v1.0.0
func (r *FileSequenceAggregator) GetInputNumber(executor.Port) int
func (*FileSequenceAggregator) GetInputs ¶ added in v1.0.0
func (r *FileSequenceAggregator) GetInputs() executor.Ports
func (*FileSequenceAggregator) GetOutputNumber ¶ added in v1.0.0
func (r *FileSequenceAggregator) GetOutputNumber(executor.Port) int
func (*FileSequenceAggregator) GetOutputs ¶ added in v1.0.0
func (r *FileSequenceAggregator) GetOutputs() executor.Ports
func (*FileSequenceAggregator) GetProcessors ¶ added in v1.0.0
func (r *FileSequenceAggregator) GetProcessors()
func (*FileSequenceAggregator) IsSink ¶ added in v1.0.0
func (r *FileSequenceAggregator) IsSink() bool
func (*FileSequenceAggregator) Name ¶ added in v1.0.0
func (r *FileSequenceAggregator) Name() string
func (*FileSequenceAggregator) Release ¶ added in v1.0.0
func (r *FileSequenceAggregator) Release() error
func (*FileSequenceAggregator) SendRecord ¶ added in v1.0.0
func (r *FileSequenceAggregator) SendRecord(re *record.Record)
type FloatCountPromOp ¶ added in v1.3.0
type FloatCountPromOp struct {
BasePromOp
}
func (*FloatCountPromOp) CreateRoutine ¶ added in v1.3.0
func (c *FloatCountPromOp) CreateRoutine(p *AggParams) (Routine, error)
type FloatHeapItem ¶
type FloatHeapItem struct {
// contains filtered or unexported fields
}
func NewFloatHeapItem ¶
func NewFloatHeapItem(n int, cmpByValue, cmpByTime func(a, b *FloatPointItem) bool) *FloatHeapItem
func (*FloatHeapItem) Len ¶
func (f *FloatHeapItem) Len() int
func (*FloatHeapItem) Less ¶
func (f *FloatHeapItem) Less(i, j int) bool
func (*FloatHeapItem) Pop ¶
func (f *FloatHeapItem) Pop() interface{}
func (*FloatHeapItem) Push ¶
func (f *FloatHeapItem) Push(x interface{})
func (*FloatHeapItem) Reset ¶
func (f *FloatHeapItem) Reset()
func (*FloatHeapItem) Swap ¶
func (f *FloatHeapItem) Swap(i, j int)
type FloatIncAggMergeFunc ¶ added in v1.3.0
type FloatIncAggMergeFunc func(prevValue float64, currValue float64, prevCount, currCount int) (float64, int)
FloatIncAggMergeFunc is used to calculate the final result based on the intermediate result and function definition.
type FloatIncAggReduceFunc ¶ added in v1.3.0
type FloatIncAggReduceFunc func(times []int64, values []float64, start, end int) (int64, float64, bool)
FloatIncAggReduceFunc is used to process intermediate calculation results.
type FloatPointItem ¶
type FloatPointItem struct {
// contains filtered or unexported fields
}
func NewFloatPointItem ¶
func NewFloatPointItem(time int64, value float64) *FloatPointItem
type FloatRateMergeFunc ¶ added in v1.3.0
type FloatRateMergeFunc func(prevTime int64, currTime int64, prevValue float64, currValue float64, ts int64, pointCount int, param *ReducerParams) (float64, bool)
FloatRateMergeFunc is used to calculate the final result based on the intermediate result and function definition.
type FloatRateReduceFunc ¶ added in v1.3.0
type FloatRateReduceFunc func(times []int64, values []float64, start, end int) (int64, int64, float64, float64, bool)
FloatRateReduceFunc is used to process intermediate calculation results. eg, calculates the first and last time points for rate.
type FloatRateUpdateFunc ¶ added in v1.3.0
type FloatRateUpdateFunc func(ft1, ft2, lt1, lt2 int64, fv1, fv2, lv1, lv2 float64) (int64, int64, float64, float64)
FloatRateUpdateFunc is used to exchange intermediate calculation results.
type FloatSliceMergeFunc ¶ added in v1.3.0
type FloatSliceReduceFunc ¶ added in v1.3.0
type HybridStoreReader ¶ added in v1.2.0
type HybridStoreReader struct { executor.BaseProcessor // contains filtered or unexported fields }
func NewHybridStoreReader ¶ added in v1.2.0
func NewHybridStoreReader(plan hybridqp.QueryNode, indexInfo *executor.CSIndexInfo) *HybridStoreReader
func (*HybridStoreReader) Abort ¶ added in v1.2.0
func (r *HybridStoreReader) Abort()
func (*HybridStoreReader) Close ¶ added in v1.2.0
func (r *HybridStoreReader) Close()
func (*HybridStoreReader) CreateCursors ¶ added in v1.2.0
func (r *HybridStoreReader) CreateCursors() error
func (*HybridStoreReader) Explain ¶ added in v1.2.0
func (r *HybridStoreReader) Explain() []executor.ValuePair
func (*HybridStoreReader) GetInputNumber ¶ added in v1.2.0
func (r *HybridStoreReader) GetInputNumber(_ executor.Port) int
func (*HybridStoreReader) GetInputs ¶ added in v1.2.0
func (r *HybridStoreReader) GetInputs() executor.Ports
func (*HybridStoreReader) GetOutputNumber ¶ added in v1.2.0
func (r *HybridStoreReader) GetOutputNumber(_ executor.Port) int
func (*HybridStoreReader) GetOutputs ¶ added in v1.2.0
func (r *HybridStoreReader) GetOutputs() executor.Ports
func (*HybridStoreReader) IsSink ¶ added in v1.2.0
func (r *HybridStoreReader) IsSink() bool
func (*HybridStoreReader) Name ¶ added in v1.2.0
func (r *HybridStoreReader) Name() string
func (*HybridStoreReader) Release ¶ added in v1.2.0
func (r *HybridStoreReader) Release() error
type IndexReader ¶ added in v1.2.0
type InstantVectorCursor ¶ added in v1.3.0
type InstantVectorCursor struct {
// contains filtered or unexported fields
}
InstantVectorCursor is used to sample and process data for prom instant_query or range_query. Sampling is the first step of original data processing for non-aggregated query and aggregated query. For non-aggregated query, place InstantVectorCursor after seriesCursor. For aggregated query, replace aggregateCursor with the InstantVectorCursor. Give an example: data: (time, value)=> [(1, 1.0), (2, 2.0), (3, 3.0), (4, 4.0), (6, 6.0)] start=1, end=6, offset=0, step=2, LookUpDelta=3 => startSample=1, endSample=5 sample data: (time, value)=> [(1, 1.0), (3, 3.0), (5, 4.0)]
func NewInstantVectorCursor ¶ added in v1.3.0
func NewInstantVectorCursor(input comm.KeyCursor, schema *executor.QuerySchema, globalPool *record.RecordPool, tr util.TimeRange) *InstantVectorCursor
func (*InstantVectorCursor) Name ¶ added in v1.3.0
func (c *InstantVectorCursor) Name() string
func (*InstantVectorCursor) Next ¶ added in v1.3.0
func (c *InstantVectorCursor) Next() (*record.Record, comm.SeriesInfoIntf, error)
func (*InstantVectorCursor) NextAggData ¶ added in v1.3.0
func (*InstantVectorCursor) SetOps ¶ added in v1.3.0
func (c *InstantVectorCursor) SetOps(ops []*comm.CallOption)
func (*InstantVectorCursor) SetSchema ¶ added in v1.3.0
func (c *InstantVectorCursor) SetSchema(inSchema, outSchema record.Schemas, exprOpt []hybridqp.ExprOptions)
func (*InstantVectorCursor) SinkPlan ¶ added in v1.3.0
func (c *InstantVectorCursor) SinkPlan(plan hybridqp.QueryNode)
type IntegerHeapItem ¶
type IntegerHeapItem struct {
// contains filtered or unexported fields
}
func NewIntegerHeapItem ¶
func NewIntegerHeapItem(n int, cmpByValue, cmpByTime func(a, b *IntegerPointItem) bool) *IntegerHeapItem
func (*IntegerHeapItem) Len ¶
func (f *IntegerHeapItem) Len() int
func (*IntegerHeapItem) Less ¶
func (f *IntegerHeapItem) Less(i, j int) bool
func (*IntegerHeapItem) Pop ¶
func (f *IntegerHeapItem) Pop() interface{}
func (*IntegerHeapItem) Push ¶
func (f *IntegerHeapItem) Push(x interface{})
func (*IntegerHeapItem) Reset ¶
func (f *IntegerHeapItem) Reset()
func (*IntegerHeapItem) Swap ¶
func (f *IntegerHeapItem) Swap(i, j int)
type IntegerPointItem ¶
type IntegerPointItem struct {
// contains filtered or unexported fields
}
func NewIntegerPointItem ¶
func NewIntegerPointItem(time int64, value int64) *IntegerPointItem
type LogReplay ¶ added in v1.0.1
type LogReplay struct {
// contains filtered or unexported fields
}
type LogReplays ¶ added in v1.0.1
type LogReplays []LogReplay
type LogWriters ¶ added in v1.0.1
type LogWriters []LogWriter
type MaxPromOp ¶ added in v1.3.0
type MaxPromOp struct {
BasePromOp
}
type MetaIndexIterator ¶ added in v1.0.0
type MetaIndexIterator struct {
// contains filtered or unexported fields
}
func NewMetaIndexIterators ¶ added in v1.0.0
func NewMetaIndexIterators(file immutable.TSSPFile, querySchema *executor.QuerySchema) (*MetaIndexIterator, error)
type MinPromOp ¶ added in v1.3.0
type MinPromOp struct {
BasePromOp
}
type NodeInfo ¶ added in v1.3.0
type NodeInfo struct {
// contains filtered or unexported fields
}
func (*NodeInfo) GenShardDirName ¶ added in v1.3.0
func (n *NodeInfo) GenShardDirName()
type PreAggTagSetCursor ¶
type PreAggTagSetCursor struct {
// contains filtered or unexported fields
}
func NewPreAggTagSetCursor ¶
func NewPreAggTagSetCursor(schema *executor.QuerySchema, ctx *idKeyCursorContext, itr comm.KeyCursor) *PreAggTagSetCursor
func (*PreAggTagSetCursor) Close ¶
func (s *PreAggTagSetCursor) Close() error
func (*PreAggTagSetCursor) EndSpan ¶
func (s *PreAggTagSetCursor) EndSpan()
func (*PreAggTagSetCursor) GetSchema ¶
func (s *PreAggTagSetCursor) GetSchema() record.Schemas
func (*PreAggTagSetCursor) Name ¶
func (s *PreAggTagSetCursor) Name() string
func (*PreAggTagSetCursor) Next ¶
func (s *PreAggTagSetCursor) Next() (*record.Record, comm.SeriesInfoIntf, error)
func (*PreAggTagSetCursor) NextAggData ¶
func (*PreAggTagSetCursor) RecordInitPreAgg ¶
func (s *PreAggTagSetCursor) RecordInitPreAgg() error
func (*PreAggTagSetCursor) SetOps ¶
func (s *PreAggTagSetCursor) SetOps(ops []*comm.CallOption)
func (*PreAggTagSetCursor) SetSchema ¶
func (s *PreAggTagSetCursor) SetSchema(schema record.Schemas)
func (*PreAggTagSetCursor) SinkPlan ¶
func (s *PreAggTagSetCursor) SinkPlan(plan hybridqp.QueryNode)
func (*PreAggTagSetCursor) StartSpan ¶
func (s *PreAggTagSetCursor) StartSpan(span *tracing.Span)
type PredictLinearOp ¶ added in v1.3.0
type PredictLinearOp struct{}
func (*PredictLinearOp) CreateRoutine ¶ added in v1.3.0
func (o *PredictLinearOp) CreateRoutine(p *PromFuncParam) (Routine, error)
type PromFuncParam ¶ added in v1.3.0
type PromFuncParam struct {
// contains filtered or unexported fields
}
type PromFunction ¶ added in v1.3.0
type PromFunction interface {
CreateRoutine(param *PromFuncParam) (Routine, error)
}
func GetPromFunction ¶ added in v1.3.0
func GetPromFunction(name string) PromFunction
type PromFunctionFactory ¶ added in v1.3.0
type PromFunctionFactory map[string]PromFunction
type QuantileOverTime ¶ added in v1.3.0
type QuantileOverTime struct{}
func (*QuantileOverTime) CreateRoutine ¶ added in v1.3.0
func (r *QuantileOverTime) CreateRoutine(param *PromFuncParam) (Routine, error)
type RangeVectorCursor ¶ added in v1.3.0
type RangeVectorCursor struct {
// contains filtered or unexported fields
}
RangeVectorCursor is used to process the calculation of the function with range duration for prom instant_query or range_query. This is a kind of sliding window calculation. For aggregated query, replace aggregateCursor with the RangeVectorCursor. Give an example: avg_over_time(value[3]) start=1, end=6, step=2 rangDuration=3, startSample=1, endSample=5 original data: (time, value)=> [(1, 1.0), (2, 2.0), (3, 3.0), (4, 4.0), (6, 6.0)] interval index: [start, end) => [[0, 1), [0,3), [1, 4)] grouped data: (time, value)=> [[(1, 1.0)], [(1, 1.0), (2, 2.0), (3, 3.0)], [(2, 2.0), (3, 3.0), (4, 4.0)]] aggregated data: (time, value)=> [ (1, 1.0), (3, 2.0), (5, 3.0)]
func NewRangeVectorCursor ¶ added in v1.3.0
func NewRangeVectorCursor(input comm.KeyCursor, schema *executor.QuerySchema, globalPool *record.RecordPool, tr util.TimeRange) *RangeVectorCursor
func (*RangeVectorCursor) Name ¶ added in v1.3.0
func (c *RangeVectorCursor) Name() string
func (*RangeVectorCursor) Next ¶ added in v1.3.0
func (c *RangeVectorCursor) Next() (*record.Record, comm.SeriesInfoIntf, error)
func (*RangeVectorCursor) NextAggData ¶ added in v1.3.0
func (*RangeVectorCursor) SetOps ¶ added in v1.3.0
func (c *RangeVectorCursor) SetOps(ops []*comm.CallOption)
func (*RangeVectorCursor) SetSchema ¶ added in v1.3.0
func (c *RangeVectorCursor) SetSchema(inSchema, outSchema record.Schemas, exprOpt []hybridqp.ExprOptions)
func (*RangeVectorCursor) SinkPlan ¶ added in v1.3.0
func (c *RangeVectorCursor) SinkPlan(plan hybridqp.QueryNode)
type Reducer ¶
type Reducer interface {
Aggregate(*ReducerEndpoint, *ReducerParams)
}
type ReducerEndpoint ¶
type ReducerEndpoint struct { InputPoint EndPointPair OutputPoint EndPointPair }
type ReducerParams ¶
type ReducerParams struct {
// contains filtered or unexported fields
}
type Routine ¶
type Routine interface {
WorkOnRecord(*record.Record, *record.Record, *ReducerParams)
}
type RoutineImpl ¶
type RoutineImpl struct {
// contains filtered or unexported fields
}
func NewRoutineImpl ¶
func NewRoutineImpl(reducer Reducer, inOrdinal int, outOrdinal int) *RoutineImpl
func (*RoutineImpl) WorkOnRecord ¶
func (r *RoutineImpl) WorkOnRecord(in *record.Record, out *record.Record, params *ReducerParams)
type SeriesIter ¶
type SeriesIter struct {
// contains filtered or unexported fields
}
type Shard ¶
type Shard interface { // IO interface WriteRows(rows []influx.Row, binaryRows []byte) error // line protocol WriteCols(mst string, cols *record.Record, binaryCols []byte) error // native protocol ForceFlush() WaitWriteFinish() CreateLogicalPlan(ctx context.Context, sources influxql.Sources, schema *executor.QuerySchema) (hybridqp.QueryNode, error) CreateCursor(ctx context.Context, schema *executor.QuerySchema) ([]comm.KeyCursor, error) Scan(span *tracing.Span, schema *executor.QuerySchema, callBack func(num int64) error) (tsi.GroupSeries, int64, error) ScanWithSparseIndex(ctx context.Context, schema *executor.QuerySchema, callBack func(num int64) error) (*executor.FileFragments, error) GetIndexInfo(schema *executor.QuerySchema) (*executor.AttachedIndexInfo, error) RowCount(schema *executor.QuerySchema) (int64, error) NewShardKeyIdx(shardType, dataPath string, lockPath *string) error // admin OpenAndEnable(client metaclient.MetaClient) error IsOpened() bool Close() error ChangeShardTierToWarm() DropMeasurement(ctx context.Context, name string) error GetSplitPoints(idxes []int64) ([]string, error) // only work for tsstore (depends on sid) // get private member GetDataPath() string GetObsOption() *obs.ObsOptions GetWalPath() string GetDuration() *meta.DurationDescriptor GetEngineType() config.EngineType GetIdent() *meta.ShardIdentifier GetID() uint64 GetRowCount() uint64 GetRPName() string GetStatistics(buffer []byte) ([]byte, error) GetMaxTime() int64 Intersect(tr *influxql.TimeRange) bool GetIndexBuilder() *tsi.IndexBuilder // only work for tsstore(tsi) GetSeriesCount() int // only work for tsstore GetTableStore() immutable.TablesStore // used by downsample and test GetTSSPFiles(mm string, isOrder bool) (*immutable.TSSPFiles, bool) // used by downsample and test GetTier() uint64 IsExpired() bool IsTierExpired() bool // downsample, only work for tsstore CanDoDownSample() bool DisableDownSample() EnableDownSample() GetShardDownSamplePolicy(policy *meta.DownSamplePolicyInfo) *meta.ShardDownSamplePolicyInfo IsOutOfOrderFilesExist() bool NewDownSampleTask(sdsp *meta.ShardDownSamplePolicyInfo, schema []hybridqp.Catalog, log *zap.Logger) SetShardDownSampleLevel(i int) SetMstInfo(mstsInfo *meta.MeasurementInfo) SetObsOption(option *obs.ObsOptions) StartDownSample(taskID uint64, level int, sdsp *meta.ShardDownSamplePolicyInfo, meta interface { UpdateShardDownSampleInfo(Ident *meta.ShardIdentifier) error }) error UpdateDownSampleOnShard(id uint64, level int) UpdateShardReadOnly(meta interface { UpdateShardDownSampleInfo(Ident *meta.ShardIdentifier) error }) error // compaction && merge, only work for tsstore Compact() error DisableCompAndMerge() EnableCompAndMerge() SetLockPath(lock *string) IsColdShard() bool CanDoShardMove() bool UnregisterShard() ExecShardMove() error DisableHierarchicalStorage() SetEnableHierarchicalStorage() CreateShowTagValuesPlan(client metaclient.MetaClient) immutable.ShowTagValuesPlan // raft SnapShot SetSnapShotter(snp *raftlog.SnapShotter) }
type ShardMoveFileInfo ¶ added in v1.3.0
type ShardStatus ¶ added in v1.1.0
func (ShardStatus) MarshalText ¶ added in v1.1.0
func (s ShardStatus) MarshalText() (data []byte, err error)
MarshalText keeps marshaled dict items order
type ShowTagValuesPlan ¶ added in v1.3.0
type ShowTagValuesPlan struct {
// contains filtered or unexported fields
}
func (*ShowTagValuesPlan) AddPlan ¶ added in v1.3.0
func (p *ShowTagValuesPlan) AddPlan(plan immutable.ShowTagValuesPlan)
func (*ShowTagValuesPlan) Execute ¶ added in v1.3.0
func (p *ShowTagValuesPlan) Execute(tagKeys map[string][][]byte, condition influxql.Expr, tr util.TimeRange, limit int) (netstorage.TablesTagSets, error)
func (*ShowTagValuesPlan) Stop ¶ added in v1.3.0
func (p *ShowTagValuesPlan) Stop()
type StdDevOverTime ¶ added in v1.3.0
type StdDevOverTime struct{}
func (*StdDevOverTime) CreateRoutine ¶ added in v1.3.0
func (r *StdDevOverTime) CreateRoutine(param *PromFuncParam) (Routine, error)
type StdVarOverTime ¶ added in v1.3.0
type StdVarOverTime struct{}
func (*StdVarOverTime) CreateRoutine ¶ added in v1.3.0
func (r *StdVarOverTime) CreateRoutine(param *PromFuncParam) (Routine, error)
type Storage ¶ added in v1.1.0
type Storage interface { WriteRowsToTable(s *shard, rows influx.Rows, mw *mstWriteCtx, binaryRows []byte) error WriteRows(s *shard, mw *mstWriteCtx) error // line protocol WriteCols(s *shard, cols *record.Record, mst string, binaryCols []byte) error // native protocol WriteIndex(s *shard, rows *influx.Rows, mw *mstWriteCtx) error SetClient(client metaclient.MetaClient) SetMstInfo(s *shard, name string, mstInfo *meta.MeasurementInfo) SetAccumulateMetaIndex(name string, detachedMetaInfo *immutable.AccumulateMetaIndex) ForceFlush(s *shard) // contains filtered or unexported methods }
type StreamDetachedReader ¶ added in v1.3.0
type StreamDetachedReader struct {
// contains filtered or unexported fields
}
StreamDetachedReader implement comm.KeyCursor and comm.TimeCutKeyCursor, it can stream read detached data to reduce IO of BF.
func NewStreamDetachedReader ¶ added in v1.3.0
func NewStreamDetachedReader(readerCtx *immutable.FileReaderContext, path *sparseindex.OBSFilterPath, ctx *indexContext) *StreamDetachedReader
func (*StreamDetachedReader) Close ¶ added in v1.3.0
func (t *StreamDetachedReader) Close() error
func (*StreamDetachedReader) EndSpan ¶ added in v1.3.0
func (t *StreamDetachedReader) EndSpan()
func (*StreamDetachedReader) GetSchema ¶ added in v1.3.0
func (t *StreamDetachedReader) GetSchema() record.Schemas
func (*StreamDetachedReader) Init ¶ added in v1.3.0
func (r *StreamDetachedReader) Init() (err error)
func (*StreamDetachedReader) Name ¶ added in v1.3.0
func (t *StreamDetachedReader) Name() string
func (*StreamDetachedReader) Next ¶ added in v1.3.0
func (t *StreamDetachedReader) Next() (*record.Record, comm.SeriesInfoIntf, error)
func (*StreamDetachedReader) NextAggData ¶ added in v1.3.0
func (*StreamDetachedReader) SetOps ¶ added in v1.3.0
func (t *StreamDetachedReader) SetOps(ops []*comm.CallOption)
func (*StreamDetachedReader) SinkPlan ¶ added in v1.3.0
func (t *StreamDetachedReader) SinkPlan(plan hybridqp.QueryNode)
func (*StreamDetachedReader) StartSpan ¶ added in v1.3.0
func (t *StreamDetachedReader) StartSpan(span *tracing.Span)
func (*StreamDetachedReader) UpdateTime ¶ added in v1.3.0
func (t *StreamDetachedReader) UpdateTime(time int64)
type TagSetCursorItem ¶
type TagSetCursorItem struct {
// contains filtered or unexported fields
}
func (TagSetCursorItem) GetNewRecord ¶
func (c TagSetCursorItem) GetNewRecord() (*record.Record, error)
type TsspSequenceReader ¶ added in v1.0.0
type TsspSequenceReader struct { executor.BaseProcessor Output *executor.SeriesRecordPort // contains filtered or unexported fields }
func (*TsspSequenceReader) Close ¶ added in v1.0.0
func (r *TsspSequenceReader) Close()
func (*TsspSequenceReader) Create ¶ added in v1.0.0
func (r *TsspSequenceReader) Create(plan executor.LogicalPlan, opt *query.ProcessorOptions) (executor.Processor, error)
func (*TsspSequenceReader) Explain ¶ added in v1.0.0
func (r *TsspSequenceReader) Explain() []executor.ValuePair
func (*TsspSequenceReader) GetInputNumber ¶ added in v1.0.0
func (r *TsspSequenceReader) GetInputNumber(executor.Port) int
func (*TsspSequenceReader) GetInputs ¶ added in v1.0.0
func (r *TsspSequenceReader) GetInputs() executor.Ports
func (*TsspSequenceReader) GetOutputNumber ¶ added in v1.0.0
func (r *TsspSequenceReader) GetOutputNumber(executor.Port) int
func (*TsspSequenceReader) GetOutputs ¶ added in v1.0.0
func (r *TsspSequenceReader) GetOutputs() executor.Ports
func (*TsspSequenceReader) IsSink ¶ added in v1.0.0
func (r *TsspSequenceReader) IsSink() bool
func (*TsspSequenceReader) Name ¶ added in v1.0.0
func (r *TsspSequenceReader) Name() string
func (*TsspSequenceReader) Release ¶ added in v1.0.0
func (r *TsspSequenceReader) Release() error
type WAL ¶
type WAL struct {
// contains filtered or unexported fields
}
type WalRecordType ¶
type WalRecordType byte
type WriteIntoStorageTransform ¶ added in v1.0.0
type WriteIntoStorageTransform struct { executor.BaseProcessor Input *executor.SeriesRecordPort Output *executor.DownSampleStatePort // contains filtered or unexported fields }
func (*WriteIntoStorageTransform) Close ¶ added in v1.0.0
func (r *WriteIntoStorageTransform) Close()
func (*WriteIntoStorageTransform) Create ¶ added in v1.0.0
func (r *WriteIntoStorageTransform) Create(plan executor.LogicalPlan, opt *query.ProcessorOptions) (executor.Processor, error)
func (*WriteIntoStorageTransform) EndFile ¶ added in v1.0.0
func (r *WriteIntoStorageTransform) EndFile() error
func (*WriteIntoStorageTransform) Explain ¶ added in v1.0.0
func (r *WriteIntoStorageTransform) Explain() []executor.ValuePair
func (*WriteIntoStorageTransform) GetClosed ¶ added in v1.0.0
func (r *WriteIntoStorageTransform) GetClosed() chan struct{}
func (*WriteIntoStorageTransform) GetInputNumber ¶ added in v1.0.0
func (r *WriteIntoStorageTransform) GetInputNumber(executor.Port) int
func (*WriteIntoStorageTransform) GetInputs ¶ added in v1.0.0
func (r *WriteIntoStorageTransform) GetInputs() executor.Ports
func (*WriteIntoStorageTransform) GetOutputNumber ¶ added in v1.0.0
func (r *WriteIntoStorageTransform) GetOutputNumber(executor.Port) int
func (*WriteIntoStorageTransform) GetOutputs ¶ added in v1.0.0
func (r *WriteIntoStorageTransform) GetOutputs() executor.Ports
func (*WriteIntoStorageTransform) GetRowCount ¶ added in v1.0.0
func (r *WriteIntoStorageTransform) GetRowCount() int
func (*WriteIntoStorageTransform) InitFile ¶ added in v1.1.0
func (r *WriteIntoStorageTransform) InitFile(sRecord *executor.SeriesRecord) error
func (*WriteIntoStorageTransform) Name ¶ added in v1.0.0
func (r *WriteIntoStorageTransform) Name() string
func (*WriteIntoStorageTransform) Release ¶ added in v1.0.0
func (r *WriteIntoStorageTransform) Release() error
func (*WriteIntoStorageTransform) SetTaskId ¶ added in v1.0.0
func (r *WriteIntoStorageTransform) SetTaskId(taskID int)
Source Files
¶
- agg_tagset_cursor.go
- aggregate_cursor.go
- backup.go
- column_store_reader.go
- compact.go
- coprocessor.go
- cs_storage.go
- detached_check_metaInfo.go
- detached_lazy_load_index_reader.go
- downsample_info.go
- engine.go
- engine_ddl.go
- engine_downsample.go
- engine_ha.go
- engine_replication.go
- fence.go
- file_cursor.go
- group_cursor.go
- hybrid_index_reader.go
- hybrid_store_reader.go
- iterator_plan.go
- iterators.go
- iterators_helper.go
- limit_cursor.go
- log_writer.go
- partition.go
- partition_raft.go
- prom_functions.go
- prom_instant_vector_cursor.go
- prom_range_vector_cursor.go
- reader_schema.go
- record_plan.go
- series_agg_func.gen.go
- series_agg_func_prom.go
- series_agg_operators.go
- series_agg_reducer.gen.go
- series_call_processor.go
- series_cursor.go
- series_loop_cursor.go
- shard.go
- shard_ddl.go
- shard_move_file_info.go
- summary.go
- sysctrl.go
- tagset_cursor.go
- tier.go
- topn_linkedlist.go
- ts_storage.go
- tsm_merge_cursor.go
- wal.go