inference

package
v2.12.0 Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Dec 8, 2023 License: BSD-3-Clause Imports: 5 Imported by: 0

Documentation

Index

Constants

This section is empty.

Variables

View Source
var File_tensorflow_core_grappler_optimizers_inference_batch_op_rewriter_proto protoreflect.FileDescriptor

Functions

This section is empty.

Types

type BatchOpRewriteConfig

type BatchOpRewriteConfig struct {
	EnableAdaptiveSharedBatchingThreadPool bool `` /* 184-byte string literal not displayed */
	// Keyed by model name, meaning all batch-ops in one saved model would use the
	// same adaptive-batch-scheduler option.
	ModelSchedulerOptions map[string]*BatchOpRewriteConfig_AdaptiveBatchSchedulerOption `` /* 214-byte string literal not displayed */
	// contains filtered or unexported fields
}

Config for the batch op rewriter. This should be serialized and set a param in RewriterConfig with key kBatchOpRewriteParamKey.

func (*BatchOpRewriteConfig) Descriptor deprecated

func (*BatchOpRewriteConfig) Descriptor() ([]byte, []int)

Deprecated: Use BatchOpRewriteConfig.ProtoReflect.Descriptor instead.

func (*BatchOpRewriteConfig) GetEnableAdaptiveSharedBatchingThreadPool

func (x *BatchOpRewriteConfig) GetEnableAdaptiveSharedBatchingThreadPool() bool

func (*BatchOpRewriteConfig) GetModelSchedulerOptions

func (*BatchOpRewriteConfig) ProtoMessage

func (*BatchOpRewriteConfig) ProtoMessage()

func (*BatchOpRewriteConfig) ProtoReflect

func (x *BatchOpRewriteConfig) ProtoReflect() protoreflect.Message

func (*BatchOpRewriteConfig) Reset

func (x *BatchOpRewriteConfig) Reset()

func (*BatchOpRewriteConfig) String

func (x *BatchOpRewriteConfig) String() string

type BatchOpRewriteConfig_AdaptiveBatchSchedulerOption

type BatchOpRewriteConfig_AdaptiveBatchSchedulerOption struct {
	MinInflightBatchesLimit     *wrapperspb.UInt32Value `` /* 134-byte string literal not displayed */
	InitialInflightBatchesLimit *wrapperspb.UInt32Value `` /* 146-byte string literal not displayed */
	MaxInflightBatchesLimit     *wrapperspb.UInt32Value `` /* 134-byte string literal not displayed */
	// You can use QPS as a reference to decide how quickly to react to workload
	// changes.
	BatchesToAverageOver *wrapperspb.UInt32Value `protobuf:"bytes,4,opt,name=batches_to_average_over,json=batchesToAverageOver,proto3" json:"batches_to_average_over,omitempty"`
	// contains filtered or unexported fields
}

The options for tensorflow::serving::AdaptiveSharedBatchScheduler. See AdaptiveSharedBatchScheduler::Options for meaning of each field.

NOTE: Leave this unset to pick up default settings which should work for most scenarios.

Example scenarios when tuning helps: * Latency sensitive

func (*BatchOpRewriteConfig_AdaptiveBatchSchedulerOption) Descriptor deprecated

Deprecated: Use BatchOpRewriteConfig_AdaptiveBatchSchedulerOption.ProtoReflect.Descriptor instead.

func (*BatchOpRewriteConfig_AdaptiveBatchSchedulerOption) GetBatchesToAverageOver

func (*BatchOpRewriteConfig_AdaptiveBatchSchedulerOption) GetInitialInflightBatchesLimit

func (x *BatchOpRewriteConfig_AdaptiveBatchSchedulerOption) GetInitialInflightBatchesLimit() *wrapperspb.UInt32Value

func (*BatchOpRewriteConfig_AdaptiveBatchSchedulerOption) GetMaxInflightBatchesLimit

func (*BatchOpRewriteConfig_AdaptiveBatchSchedulerOption) GetMinInflightBatchesLimit

func (*BatchOpRewriteConfig_AdaptiveBatchSchedulerOption) ProtoMessage

func (*BatchOpRewriteConfig_AdaptiveBatchSchedulerOption) ProtoReflect

func (*BatchOpRewriteConfig_AdaptiveBatchSchedulerOption) Reset

func (*BatchOpRewriteConfig_AdaptiveBatchSchedulerOption) String

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL