llm

package
v2.2.0 Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Jan 30, 2024 License: Apache-2.0 Imports: 4 Imported by: 0

Documentation

Overview

Package llm provides the interface to use Large Language Models in Spin.

Index

Constants

This section is empty.

Variables

This section is empty.

Functions

This section is empty.

Types

type EmbeddingsResult

type EmbeddingsResult struct {
	// Embeddings are the embeddings generated by the request.
	Embeddings [][]float32
	// Usage is usage related to an embeddings generation request.
	Usage *EmbeddingsUsage
}

EmbeddingsResult of generating embeddings.

func GenerateEmbeddings

func GenerateEmbeddings(model string, text []string) (*EmbeddingsResult, error)

GenerateEmbeddings generates the embeddings for the supplied list of text.

type EmbeddingsUsage

type EmbeddingsUsage struct {
	// PromptTokenCount is number of tokens in the prompt.
	PromptTokenCount int
}

Embeddings is usage related to an embeddings generation request.

type InferencingParams

type InferencingParams struct {
	// MaxTokens is the maximum tokens that should be inferred.
	// Default: 100
	//
	// Note: the backing implementation may return less tokens.
	MaxTokens int32
	// RepeatPenalty is the amount the model should avoid repeating tokens.
	// Default: 1.1
	RepeatPenalty float32
	// RepeatPenaltyLastNTokenCount the number of tokens the model should
	// apply the repeat penalty to.
	// Default: 64
	RepeatPenaltyLastNTokenCount int32
	// Temperature is the randomness with which the next token is selected.
	// Default: 0.8
	Temperature float32
	// TopK is the number of possible next tokens the model will choose from.
	// Default: 40
	TopK int32
	// TopP is the probability total of next tokens the model will choose
	// from.
	// Default: 0.9
	TopP float32
}

InferenceParams is the optional request parameters.

type InferencingResult

type InferencingResult struct {
	// Text is the text generated by the model.
	Text string
	// Usage is information about the inferencing request.
	Usage *InferencingUsage
}

InferencingResult is the result of an inference.

func Infer

func Infer(model, prompt string, params *InferencingParams) (*InferencingResult, error)

Infer performs inferencing using the provided model and prompt with the given optional parameters.

type InferencingUsage

type InferencingUsage struct {
	// PromptTokenCount is the number of tokens in the prompt.
	PromptTokenCount int
	// GeneratedTokenCount is the number of tokens generated by the
	// inferencing operation.
	GeneratedTokenCount int
}

InferencingUsage represents information related to the inferencing result.

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL