Documentation ¶
Overview ¶
Package speech is a generated protocol buffer package.
It is generated from these files:
google/cloud/speech/v1p1beta1/cloud_speech.proto
It has these top-level messages:
RecognizeRequest LongRunningRecognizeRequest StreamingRecognizeRequest StreamingRecognitionConfig RecognitionConfig RecognitionMetadata SpeechContext RecognitionAudio RecognizeResponse LongRunningRecognizeResponse LongRunningRecognizeMetadata StreamingRecognizeResponse StreamingRecognitionResult SpeechRecognitionResult SpeechRecognitionAlternative WordInfo
Index ¶
- Variables
- func RegisterSpeechServer(s *grpc.Server, srv SpeechServer)
- type LongRunningRecognizeMetadata
- func (*LongRunningRecognizeMetadata) Descriptor() ([]byte, []int)
- func (m *LongRunningRecognizeMetadata) GetLastUpdateTime() *google_protobuf4.Timestamp
- func (m *LongRunningRecognizeMetadata) GetProgressPercent() int32
- func (m *LongRunningRecognizeMetadata) GetStartTime() *google_protobuf4.Timestamp
- func (*LongRunningRecognizeMetadata) ProtoMessage()
- func (m *LongRunningRecognizeMetadata) Reset()
- func (m *LongRunningRecognizeMetadata) String() string
- type LongRunningRecognizeRequest
- func (*LongRunningRecognizeRequest) Descriptor() ([]byte, []int)
- func (m *LongRunningRecognizeRequest) GetAudio() *RecognitionAudio
- func (m *LongRunningRecognizeRequest) GetConfig() *RecognitionConfig
- func (*LongRunningRecognizeRequest) ProtoMessage()
- func (m *LongRunningRecognizeRequest) Reset()
- func (m *LongRunningRecognizeRequest) String() string
- type LongRunningRecognizeResponse
- type RecognitionAudio
- func (*RecognitionAudio) Descriptor() ([]byte, []int)
- func (m *RecognitionAudio) GetAudioSource() isRecognitionAudio_AudioSource
- func (m *RecognitionAudio) GetContent() []byte
- func (m *RecognitionAudio) GetUri() string
- func (*RecognitionAudio) ProtoMessage()
- func (m *RecognitionAudio) Reset()
- func (m *RecognitionAudio) String() string
- func (*RecognitionAudio) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, ...)
- type RecognitionAudio_Content
- type RecognitionAudio_Uri
- type RecognitionConfig
- func (*RecognitionConfig) Descriptor() ([]byte, []int)
- func (m *RecognitionConfig) GetEnableAutomaticPunctuation() bool
- func (m *RecognitionConfig) GetEnableWordTimeOffsets() bool
- func (m *RecognitionConfig) GetEncoding() RecognitionConfig_AudioEncoding
- func (m *RecognitionConfig) GetLanguageCode() string
- func (m *RecognitionConfig) GetMaxAlternatives() int32
- func (m *RecognitionConfig) GetMetadata() *RecognitionMetadata
- func (m *RecognitionConfig) GetModel() string
- func (m *RecognitionConfig) GetProfanityFilter() bool
- func (m *RecognitionConfig) GetSampleRateHertz() int32
- func (m *RecognitionConfig) GetSpeechContexts() []*SpeechContext
- func (m *RecognitionConfig) GetUseEnhanced() bool
- func (*RecognitionConfig) ProtoMessage()
- func (m *RecognitionConfig) Reset()
- func (m *RecognitionConfig) String() string
- type RecognitionConfig_AudioEncoding
- type RecognitionMetadata
- func (*RecognitionMetadata) Descriptor() ([]byte, []int)
- func (m *RecognitionMetadata) GetAudioTopic() string
- func (m *RecognitionMetadata) GetIndustryNaicsCodeOfAudio() uint32
- func (m *RecognitionMetadata) GetInteractionType() RecognitionMetadata_InteractionType
- func (m *RecognitionMetadata) GetMicrophoneDistance() RecognitionMetadata_MicrophoneDistance
- func (m *RecognitionMetadata) GetObfuscatedId() int64
- func (m *RecognitionMetadata) GetOriginalMediaType() RecognitionMetadata_OriginalMediaType
- func (m *RecognitionMetadata) GetOriginalMimeType() string
- func (m *RecognitionMetadata) GetRecordingDeviceName() string
- func (m *RecognitionMetadata) GetRecordingDeviceType() RecognitionMetadata_RecordingDeviceType
- func (*RecognitionMetadata) ProtoMessage()
- func (m *RecognitionMetadata) Reset()
- func (m *RecognitionMetadata) String() string
- type RecognitionMetadata_InteractionType
- type RecognitionMetadata_MicrophoneDistance
- type RecognitionMetadata_OriginalMediaType
- type RecognitionMetadata_RecordingDeviceType
- type RecognizeRequest
- type RecognizeResponse
- type SpeechClient
- type SpeechContext
- type SpeechRecognitionAlternative
- func (*SpeechRecognitionAlternative) Descriptor() ([]byte, []int)
- func (m *SpeechRecognitionAlternative) GetConfidence() float32
- func (m *SpeechRecognitionAlternative) GetTranscript() string
- func (m *SpeechRecognitionAlternative) GetWords() []*WordInfo
- func (*SpeechRecognitionAlternative) ProtoMessage()
- func (m *SpeechRecognitionAlternative) Reset()
- func (m *SpeechRecognitionAlternative) String() string
- type SpeechRecognitionResult
- type SpeechServer
- type Speech_StreamingRecognizeClient
- type Speech_StreamingRecognizeServer
- type StreamingRecognitionConfig
- func (*StreamingRecognitionConfig) Descriptor() ([]byte, []int)
- func (m *StreamingRecognitionConfig) GetConfig() *RecognitionConfig
- func (m *StreamingRecognitionConfig) GetInterimResults() bool
- func (m *StreamingRecognitionConfig) GetSingleUtterance() bool
- func (*StreamingRecognitionConfig) ProtoMessage()
- func (m *StreamingRecognitionConfig) Reset()
- func (m *StreamingRecognitionConfig) String() string
- type StreamingRecognitionResult
- func (*StreamingRecognitionResult) Descriptor() ([]byte, []int)
- func (m *StreamingRecognitionResult) GetAlternatives() []*SpeechRecognitionAlternative
- func (m *StreamingRecognitionResult) GetIsFinal() bool
- func (m *StreamingRecognitionResult) GetStability() float32
- func (*StreamingRecognitionResult) ProtoMessage()
- func (m *StreamingRecognitionResult) Reset()
- func (m *StreamingRecognitionResult) String() string
- type StreamingRecognizeRequest
- func (*StreamingRecognizeRequest) Descriptor() ([]byte, []int)
- func (m *StreamingRecognizeRequest) GetAudioContent() []byte
- func (m *StreamingRecognizeRequest) GetStreamingConfig() *StreamingRecognitionConfig
- func (m *StreamingRecognizeRequest) GetStreamingRequest() isStreamingRecognizeRequest_StreamingRequest
- func (*StreamingRecognizeRequest) ProtoMessage()
- func (m *StreamingRecognizeRequest) Reset()
- func (m *StreamingRecognizeRequest) String() string
- func (*StreamingRecognizeRequest) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, ...)
- type StreamingRecognizeRequest_AudioContent
- type StreamingRecognizeRequest_StreamingConfig
- type StreamingRecognizeResponse
- func (*StreamingRecognizeResponse) Descriptor() ([]byte, []int)
- func (m *StreamingRecognizeResponse) GetError() *google_rpc.Status
- func (m *StreamingRecognizeResponse) GetResults() []*StreamingRecognitionResult
- func (m *StreamingRecognizeResponse) GetSpeechEventType() StreamingRecognizeResponse_SpeechEventType
- func (*StreamingRecognizeResponse) ProtoMessage()
- func (m *StreamingRecognizeResponse) Reset()
- func (m *StreamingRecognizeResponse) String() string
- type StreamingRecognizeResponse_SpeechEventType
- type WordInfo
Constants ¶
This section is empty.
Variables ¶
var RecognitionConfig_AudioEncoding_name = map[int32]string{
0: "ENCODING_UNSPECIFIED",
1: "LINEAR16",
2: "FLAC",
3: "MULAW",
4: "AMR",
5: "AMR_WB",
6: "OGG_OPUS",
7: "SPEEX_WITH_HEADER_BYTE",
}
var RecognitionConfig_AudioEncoding_value = map[string]int32{
"ENCODING_UNSPECIFIED": 0,
"LINEAR16": 1,
"FLAC": 2,
"MULAW": 3,
"AMR": 4,
"AMR_WB": 5,
"OGG_OPUS": 6,
"SPEEX_WITH_HEADER_BYTE": 7,
}
var RecognitionMetadata_InteractionType_name = map[int32]string{
0: "INTERACTION_TYPE_UNSPECIFIED",
1: "DISCUSSION",
2: "PRESENTATION",
3: "PHONE_CALL",
4: "VOICEMAIL",
5: "PROFESSIONALLY_PRODUCED",
6: "VOICE_SEARCH",
7: "VOICE_COMMAND",
8: "DICTATION",
}
var RecognitionMetadata_InteractionType_value = map[string]int32{
"INTERACTION_TYPE_UNSPECIFIED": 0,
"DISCUSSION": 1,
"PRESENTATION": 2,
"PHONE_CALL": 3,
"VOICEMAIL": 4,
"PROFESSIONALLY_PRODUCED": 5,
"VOICE_SEARCH": 6,
"VOICE_COMMAND": 7,
"DICTATION": 8,
}
var RecognitionMetadata_MicrophoneDistance_name = map[int32]string{
0: "MICROPHONE_DISTANCE_UNSPECIFIED",
1: "NEARFIELD",
2: "MIDFIELD",
3: "FARFIELD",
}
var RecognitionMetadata_MicrophoneDistance_value = map[string]int32{
"MICROPHONE_DISTANCE_UNSPECIFIED": 0,
"NEARFIELD": 1,
"MIDFIELD": 2,
"FARFIELD": 3,
}
var RecognitionMetadata_OriginalMediaType_name = map[int32]string{
0: "ORIGINAL_MEDIA_TYPE_UNSPECIFIED",
1: "AUDIO",
2: "VIDEO",
}
var RecognitionMetadata_OriginalMediaType_value = map[string]int32{
"ORIGINAL_MEDIA_TYPE_UNSPECIFIED": 0,
"AUDIO": 1,
"VIDEO": 2,
}
var RecognitionMetadata_RecordingDeviceType_name = map[int32]string{
0: "RECORDING_DEVICE_TYPE_UNSPECIFIED",
1: "SMARTPHONE",
2: "PC",
3: "PHONE_LINE",
4: "VEHICLE",
5: "OTHER_OUTDOOR_DEVICE",
6: "OTHER_INDOOR_DEVICE",
}
var RecognitionMetadata_RecordingDeviceType_value = map[string]int32{
"RECORDING_DEVICE_TYPE_UNSPECIFIED": 0,
"SMARTPHONE": 1,
"PC": 2,
"PHONE_LINE": 3,
"VEHICLE": 4,
"OTHER_OUTDOOR_DEVICE": 5,
"OTHER_INDOOR_DEVICE": 6,
}
var StreamingRecognizeResponse_SpeechEventType_name = map[int32]string{
0: "SPEECH_EVENT_UNSPECIFIED",
1: "END_OF_SINGLE_UTTERANCE",
}
var StreamingRecognizeResponse_SpeechEventType_value = map[string]int32{
"SPEECH_EVENT_UNSPECIFIED": 0,
"END_OF_SINGLE_UTTERANCE": 1,
}
Functions ¶
func RegisterSpeechServer ¶
func RegisterSpeechServer(s *grpc.Server, srv SpeechServer)
Types ¶
type LongRunningRecognizeMetadata ¶
type LongRunningRecognizeMetadata struct { // Approximate percentage of audio processed thus far. Guaranteed to be 100 // when the audio is fully processed and the results are available. ProgressPercent int32 `protobuf:"varint,1,opt,name=progress_percent,json=progressPercent" json:"progress_percent,omitempty"` // Time when the request was received. StartTime *google_protobuf4.Timestamp `protobuf:"bytes,2,opt,name=start_time,json=startTime" json:"start_time,omitempty"` // Time of the most recent processing update. LastUpdateTime *google_protobuf4.Timestamp `protobuf:"bytes,3,opt,name=last_update_time,json=lastUpdateTime" json:"last_update_time,omitempty"` }
Describes the progress of a long-running `LongRunningRecognize` call. It is included in the `metadata` field of the `Operation` returned by the `GetOperation` call of the `google::longrunning::Operations` service.
func (*LongRunningRecognizeMetadata) Descriptor ¶
func (*LongRunningRecognizeMetadata) Descriptor() ([]byte, []int)
func (*LongRunningRecognizeMetadata) GetLastUpdateTime ¶
func (m *LongRunningRecognizeMetadata) GetLastUpdateTime() *google_protobuf4.Timestamp
func (*LongRunningRecognizeMetadata) GetProgressPercent ¶
func (m *LongRunningRecognizeMetadata) GetProgressPercent() int32
func (*LongRunningRecognizeMetadata) GetStartTime ¶
func (m *LongRunningRecognizeMetadata) GetStartTime() *google_protobuf4.Timestamp
func (*LongRunningRecognizeMetadata) ProtoMessage ¶
func (*LongRunningRecognizeMetadata) ProtoMessage()
func (*LongRunningRecognizeMetadata) Reset ¶
func (m *LongRunningRecognizeMetadata) Reset()
func (*LongRunningRecognizeMetadata) String ¶
func (m *LongRunningRecognizeMetadata) String() string
type LongRunningRecognizeRequest ¶
type LongRunningRecognizeRequest struct { // *Required* Provides information to the recognizer that specifies how to // process the request. Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config" json:"config,omitempty"` // *Required* The audio data to be recognized. Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio" json:"audio,omitempty"` }
The top-level message sent by the client for the `LongRunningRecognize` method.
func (*LongRunningRecognizeRequest) Descriptor ¶
func (*LongRunningRecognizeRequest) Descriptor() ([]byte, []int)
func (*LongRunningRecognizeRequest) GetAudio ¶
func (m *LongRunningRecognizeRequest) GetAudio() *RecognitionAudio
func (*LongRunningRecognizeRequest) GetConfig ¶
func (m *LongRunningRecognizeRequest) GetConfig() *RecognitionConfig
func (*LongRunningRecognizeRequest) ProtoMessage ¶
func (*LongRunningRecognizeRequest) ProtoMessage()
func (*LongRunningRecognizeRequest) Reset ¶
func (m *LongRunningRecognizeRequest) Reset()
func (*LongRunningRecognizeRequest) String ¶
func (m *LongRunningRecognizeRequest) String() string
type LongRunningRecognizeResponse ¶
type LongRunningRecognizeResponse struct { // Output only. Sequential list of transcription results corresponding to // sequential portions of audio. Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results" json:"results,omitempty"` }
The only message returned to the client by the `LongRunningRecognize` method. It contains the result as zero or more sequential `SpeechRecognitionResult` messages. It is included in the `result.response` field of the `Operation` returned by the `GetOperation` call of the `google::longrunning::Operations` service.
func (*LongRunningRecognizeResponse) Descriptor ¶
func (*LongRunningRecognizeResponse) Descriptor() ([]byte, []int)
func (*LongRunningRecognizeResponse) GetResults ¶
func (m *LongRunningRecognizeResponse) GetResults() []*SpeechRecognitionResult
func (*LongRunningRecognizeResponse) ProtoMessage ¶
func (*LongRunningRecognizeResponse) ProtoMessage()
func (*LongRunningRecognizeResponse) Reset ¶
func (m *LongRunningRecognizeResponse) Reset()
func (*LongRunningRecognizeResponse) String ¶
func (m *LongRunningRecognizeResponse) String() string
type RecognitionAudio ¶
type RecognitionAudio struct { // The audio source, which is either inline content or a Google Cloud // Storage uri. // // Types that are valid to be assigned to AudioSource: // *RecognitionAudio_Content // *RecognitionAudio_Uri AudioSource isRecognitionAudio_AudioSource `protobuf_oneof:"audio_source"` }
Contains audio data in the encoding specified in the `RecognitionConfig`. Either `content` or `uri` must be supplied. Supplying both or neither returns [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]. See [audio limits](https://cloud.google.com/speech/limits#content).
func (*RecognitionAudio) Descriptor ¶
func (*RecognitionAudio) Descriptor() ([]byte, []int)
func (*RecognitionAudio) GetAudioSource ¶
func (m *RecognitionAudio) GetAudioSource() isRecognitionAudio_AudioSource
func (*RecognitionAudio) GetContent ¶
func (m *RecognitionAudio) GetContent() []byte
func (*RecognitionAudio) GetUri ¶
func (m *RecognitionAudio) GetUri() string
func (*RecognitionAudio) ProtoMessage ¶
func (*RecognitionAudio) ProtoMessage()
func (*RecognitionAudio) Reset ¶
func (m *RecognitionAudio) Reset()
func (*RecognitionAudio) String ¶
func (m *RecognitionAudio) String() string
func (*RecognitionAudio) XXX_OneofFuncs ¶
func (*RecognitionAudio) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{})
XXX_OneofFuncs is for the internal use of the proto package.
type RecognitionAudio_Content ¶
type RecognitionAudio_Content struct {
Content []byte `protobuf:"bytes,1,opt,name=content,proto3,oneof"`
}
type RecognitionAudio_Uri ¶
type RecognitionAudio_Uri struct {
Uri string `protobuf:"bytes,2,opt,name=uri,oneof"`
}
type RecognitionConfig ¶
type RecognitionConfig struct { // Encoding of audio data sent in all `RecognitionAudio` messages. // This field is optional for `FLAC` and `WAV` audio files and required // for all other audio formats. For details, see [AudioEncoding][google.cloud.speech.v1p1beta1.RecognitionConfig.AudioEncoding]. Encoding RecognitionConfig_AudioEncoding `` /* 130-byte string literal not displayed */ // Sample rate in Hertz of the audio data sent in all // `RecognitionAudio` messages. Valid values are: 8000-48000. // 16000 is optimal. For best results, set the sampling rate of the audio // source to 16000 Hz. If that's not possible, use the native sample rate of // the audio source (instead of re-sampling). // This field is optional for `FLAC` and `WAV` audio files and required // for all other audio formats. For details, see [AudioEncoding][google.cloud.speech.v1p1beta1.RecognitionConfig.AudioEncoding]. SampleRateHertz int32 `protobuf:"varint,2,opt,name=sample_rate_hertz,json=sampleRateHertz" json:"sample_rate_hertz,omitempty"` // *Required* The language of the supplied audio as a // [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. // Example: "en-US". // See [Language Support](https://cloud.google.com/speech/docs/languages) // for a list of the currently supported language codes. LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"` // *Optional* Maximum number of recognition hypotheses to be returned. // Specifically, the maximum number of `SpeechRecognitionAlternative` messages // within each `SpeechRecognitionResult`. // The server may return fewer than `max_alternatives`. // Valid values are `0`-`30`. A value of `0` or `1` will return a maximum of // one. If omitted, will return a maximum of one. MaxAlternatives int32 `protobuf:"varint,4,opt,name=max_alternatives,json=maxAlternatives" json:"max_alternatives,omitempty"` // *Optional* If set to `true`, the server will attempt to filter out // profanities, replacing all but the initial character in each filtered word // with asterisks, e.g. "f***". If set to `false` or omitted, profanities // won't be filtered out. ProfanityFilter bool `protobuf:"varint,5,opt,name=profanity_filter,json=profanityFilter" json:"profanity_filter,omitempty"` // *Optional* A means to provide context to assist the speech recognition. SpeechContexts []*SpeechContext `protobuf:"bytes,6,rep,name=speech_contexts,json=speechContexts" json:"speech_contexts,omitempty"` // *Optional* If `true`, the top result includes a list of words and // the start and end time offsets (timestamps) for those words. If // `false`, no word-level time offset information is returned. The default is // `false`. EnableWordTimeOffsets bool `protobuf:"varint,8,opt,name=enable_word_time_offsets,json=enableWordTimeOffsets" json:"enable_word_time_offsets,omitempty"` // *Optional* If 'true', adds punctuation to recognition result hypotheses. // This feature is only available in select languages. Setting this for // requests in other languages has no effect at all. // The default 'false' value does not add punctuation to result hypotheses. // NOTE: "This is currently offered as an experimental service, complimentary // to all users. In the future this may be exclusively available as a // premium feature." EnableAutomaticPunctuation bool `` /* 136-byte string literal not displayed */ // *Optional* Metadata regarding this request. Metadata *RecognitionMetadata `protobuf:"bytes,9,opt,name=metadata" json:"metadata,omitempty"` // *Optional* Which model to select for the given request. Select the model // best suited to your domain to get best results. If a model is not // explicitly specified, then we auto-select a model based on the parameters // in the RecognitionConfig. // <table> // <tr> // <td><b>Model</b></td> // <td><b>Description</b></td> // </tr> // <tr> // <td><code>command_and_search</code></td> // <td>Best for short queries such as voice commands or voice search.</td> // </tr> // <tr> // <td><code>phone_call</code></td> // <td>Best for audio that originated from a phone call (typically // recorded at an 8khz sampling rate).</td> // </tr> // <tr> // <td><code>video</code></td> // <td>Best for audio that originated from from video or includes multiple // speakers. Ideally the audio is recorded at a 16khz or greater // sampling rate. This is a premium model that costs more than the // standard rate.</td> // </tr> // <tr> // <td><code>default</code></td> // <td>Best for audio that is not one of the specific audio models. // For example, long-form audio. Ideally the audio is high-fidelity, // recorded at a 16khz or greater sampling rate.</td> // </tr> // </table> Model string `protobuf:"bytes,13,opt,name=model" json:"model,omitempty"` // *Optional* Set to true to use an enhanced model for speech recognition. // You must also set the `model` field to a valid, enhanced model. If // `use_enhanced` is set to true and the `model` field is not set, then // `use_enhanced` is ignored. If `use_enhanced` is true and an enhanced // version of the specified model does not exist, then the speech is // recognized using the standard version of the specified model. // // Enhanced speech models require that you opt-in to the audio logging using // instructions in the [alpha documentation](/speech/data-sharing). If you set // `use_enhanced` to true and you have not enabled audio logging, then you // will receive an error. UseEnhanced bool `protobuf:"varint,14,opt,name=use_enhanced,json=useEnhanced" json:"use_enhanced,omitempty"` }
Provides information to the recognizer that specifies how to process the request.
func (*RecognitionConfig) Descriptor ¶
func (*RecognitionConfig) Descriptor() ([]byte, []int)
func (*RecognitionConfig) GetEnableAutomaticPunctuation ¶
func (m *RecognitionConfig) GetEnableAutomaticPunctuation() bool
func (*RecognitionConfig) GetEnableWordTimeOffsets ¶
func (m *RecognitionConfig) GetEnableWordTimeOffsets() bool
func (*RecognitionConfig) GetEncoding ¶
func (m *RecognitionConfig) GetEncoding() RecognitionConfig_AudioEncoding
func (*RecognitionConfig) GetLanguageCode ¶
func (m *RecognitionConfig) GetLanguageCode() string
func (*RecognitionConfig) GetMaxAlternatives ¶
func (m *RecognitionConfig) GetMaxAlternatives() int32
func (*RecognitionConfig) GetMetadata ¶
func (m *RecognitionConfig) GetMetadata() *RecognitionMetadata
func (*RecognitionConfig) GetModel ¶
func (m *RecognitionConfig) GetModel() string
func (*RecognitionConfig) GetProfanityFilter ¶
func (m *RecognitionConfig) GetProfanityFilter() bool
func (*RecognitionConfig) GetSampleRateHertz ¶
func (m *RecognitionConfig) GetSampleRateHertz() int32
func (*RecognitionConfig) GetSpeechContexts ¶
func (m *RecognitionConfig) GetSpeechContexts() []*SpeechContext
func (*RecognitionConfig) GetUseEnhanced ¶
func (m *RecognitionConfig) GetUseEnhanced() bool
func (*RecognitionConfig) ProtoMessage ¶
func (*RecognitionConfig) ProtoMessage()
func (*RecognitionConfig) Reset ¶
func (m *RecognitionConfig) Reset()
func (*RecognitionConfig) String ¶
func (m *RecognitionConfig) String() string
type RecognitionConfig_AudioEncoding ¶
type RecognitionConfig_AudioEncoding int32
The encoding of the audio data sent in the request.
All encodings support only 1 channel (mono) audio.
For best results, the audio source should be captured and transmitted using a lossless encoding (`FLAC` or `LINEAR16`). The accuracy of the speech recognition can be reduced if lossy codecs are used to capture or transmit audio, particularly if background noise is present. Lossy codecs include `MULAW`, `AMR`, `AMR_WB`, `OGG_OPUS`, and `SPEEX_WITH_HEADER_BYTE`.
The `FLAC` and `WAV` audio file formats include a header that describes the included audio content. You can request recognition for `WAV` files that contain either `LINEAR16` or `MULAW` encoded audio. If you send `FLAC` or `WAV` audio file format in your request, you do not need to specify an `AudioEncoding`; the audio encoding format is determined from the file header. If you specify an `AudioEncoding` when you send send `FLAC` or `WAV` audio, the encoding configuration must match the encoding described in the audio header; otherwise the request returns an [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT] error code.
const ( // Not specified. RecognitionConfig_ENCODING_UNSPECIFIED RecognitionConfig_AudioEncoding = 0 // Uncompressed 16-bit signed little-endian samples (Linear PCM). RecognitionConfig_LINEAR16 RecognitionConfig_AudioEncoding = 1 // `FLAC` (Free Lossless Audio // Codec) is the recommended encoding because it is // lossless--therefore recognition is not compromised--and // requires only about half the bandwidth of `LINEAR16`. `FLAC` stream // encoding supports 16-bit and 24-bit samples, however, not all fields in // `STREAMINFO` are supported. RecognitionConfig_FLAC RecognitionConfig_AudioEncoding = 2 // 8-bit samples that compand 14-bit audio samples using G.711 PCMU/mu-law. RecognitionConfig_MULAW RecognitionConfig_AudioEncoding = 3 // Adaptive Multi-Rate Narrowband codec. `sample_rate_hertz` must be 8000. RecognitionConfig_AMR RecognitionConfig_AudioEncoding = 4 // Adaptive Multi-Rate Wideband codec. `sample_rate_hertz` must be 16000. RecognitionConfig_AMR_WB RecognitionConfig_AudioEncoding = 5 // Opus encoded audio frames in Ogg container // ([OggOpus](https://wiki.xiph.org/OggOpus)). // `sample_rate_hertz` must be one of 8000, 12000, 16000, 24000, or 48000. RecognitionConfig_OGG_OPUS RecognitionConfig_AudioEncoding = 6 // Although the use of lossy encodings is not recommended, if a very low // bitrate encoding is required, `OGG_OPUS` is highly preferred over // Speex encoding. The [Speex](https://speex.org/) encoding supported by // Cloud Speech API has a header byte in each block, as in MIME type // `audio/x-speex-with-header-byte`. // It is a variant of the RTP Speex encoding defined in // [RFC 5574](https://tools.ietf.org/html/rfc5574). // The stream is a sequence of blocks, one block per RTP packet. Each block // starts with a byte containing the length of the block, in bytes, followed // by one or more frames of Speex data, padded to an integral number of // bytes (octets) as specified in RFC 5574. In other words, each RTP header // is replaced with a single byte containing the block length. Only Speex // wideband is supported. `sample_rate_hertz` must be 16000. RecognitionConfig_SPEEX_WITH_HEADER_BYTE RecognitionConfig_AudioEncoding = 7 )
func (RecognitionConfig_AudioEncoding) EnumDescriptor ¶
func (RecognitionConfig_AudioEncoding) EnumDescriptor() ([]byte, []int)
func (RecognitionConfig_AudioEncoding) String ¶
func (x RecognitionConfig_AudioEncoding) String() string
type RecognitionMetadata ¶
type RecognitionMetadata struct { // The use case most closely describing the audio content to be recognized. InteractionType RecognitionMetadata_InteractionType `` /* 171-byte string literal not displayed */ // The industry vertical to which this speech recognition request most // closely applies. This is most indicative of the topics contained // in the audio. Use the 6-digit NAICS code to identify the industry // vertical - see https://www.naics.com/search/. IndustryNaicsCodeOfAudio uint32 `` /* 133-byte string literal not displayed */ // The audio type that most closely describes the audio being recognized. MicrophoneDistance RecognitionMetadata_MicrophoneDistance `` /* 183-byte string literal not displayed */ // The original media the speech was recorded on. OriginalMediaType RecognitionMetadata_OriginalMediaType `` /* 181-byte string literal not displayed */ // The type of device the speech was recorded with. RecordingDeviceType RecognitionMetadata_RecordingDeviceType `` /* 189-byte string literal not displayed */ // The device used to make the recording. Examples 'Nexus 5X' or // 'Polycom SoundStation IP 6000' or 'POTS' or 'VoIP' or // 'Cardioid Microphone'. RecordingDeviceName string `protobuf:"bytes,7,opt,name=recording_device_name,json=recordingDeviceName" json:"recording_device_name,omitempty"` // Mime type of the original audio file. For example `audio/m4a`, // `audio/x-alaw-basic`, `audio/mp3`, `audio/3gpp`. // A list of possible audio mime types is maintained at // http://www.iana.org/assignments/media-types/media-types.xhtml#audio OriginalMimeType string `protobuf:"bytes,8,opt,name=original_mime_type,json=originalMimeType" json:"original_mime_type,omitempty"` // Obfuscated (privacy-protected) ID of the user, to identify number of // unique users using the service. ObfuscatedId int64 `protobuf:"varint,9,opt,name=obfuscated_id,json=obfuscatedId" json:"obfuscated_id,omitempty"` // Description of the content. Eg. "Recordings of federal supreme court // hearings from 2012". AudioTopic string `protobuf:"bytes,10,opt,name=audio_topic,json=audioTopic" json:"audio_topic,omitempty"` }
Description of audio data to be recognized.
func (*RecognitionMetadata) Descriptor ¶
func (*RecognitionMetadata) Descriptor() ([]byte, []int)
func (*RecognitionMetadata) GetAudioTopic ¶
func (m *RecognitionMetadata) GetAudioTopic() string
func (*RecognitionMetadata) GetIndustryNaicsCodeOfAudio ¶
func (m *RecognitionMetadata) GetIndustryNaicsCodeOfAudio() uint32
func (*RecognitionMetadata) GetInteractionType ¶
func (m *RecognitionMetadata) GetInteractionType() RecognitionMetadata_InteractionType
func (*RecognitionMetadata) GetMicrophoneDistance ¶
func (m *RecognitionMetadata) GetMicrophoneDistance() RecognitionMetadata_MicrophoneDistance
func (*RecognitionMetadata) GetObfuscatedId ¶
func (m *RecognitionMetadata) GetObfuscatedId() int64
func (*RecognitionMetadata) GetOriginalMediaType ¶
func (m *RecognitionMetadata) GetOriginalMediaType() RecognitionMetadata_OriginalMediaType
func (*RecognitionMetadata) GetOriginalMimeType ¶
func (m *RecognitionMetadata) GetOriginalMimeType() string
func (*RecognitionMetadata) GetRecordingDeviceName ¶
func (m *RecognitionMetadata) GetRecordingDeviceName() string
func (*RecognitionMetadata) GetRecordingDeviceType ¶
func (m *RecognitionMetadata) GetRecordingDeviceType() RecognitionMetadata_RecordingDeviceType
func (*RecognitionMetadata) ProtoMessage ¶
func (*RecognitionMetadata) ProtoMessage()
func (*RecognitionMetadata) Reset ¶
func (m *RecognitionMetadata) Reset()
func (*RecognitionMetadata) String ¶
func (m *RecognitionMetadata) String() string
type RecognitionMetadata_InteractionType ¶
type RecognitionMetadata_InteractionType int32
Use case categories that the audio recognition request can be described by.
const ( // Use case is either unknown or is something other than one of the other // values below. RecognitionMetadata_INTERACTION_TYPE_UNSPECIFIED RecognitionMetadata_InteractionType = 0 // Multiple people in a conversation or discussion. For example in a // meeting with two or more people actively participating. Typically // all the primary people speaking would be in the same room (if not, // see PHONE_CALL) RecognitionMetadata_DISCUSSION RecognitionMetadata_InteractionType = 1 // One or more persons lecturing or presenting to others, mostly // uninterrupted. RecognitionMetadata_PRESENTATION RecognitionMetadata_InteractionType = 2 // A phone-call or video-conference in which two or more people, who are // not in the same room, are actively participating. RecognitionMetadata_PHONE_CALL RecognitionMetadata_InteractionType = 3 // A recorded message intended for another person to listen to. RecognitionMetadata_VOICEMAIL RecognitionMetadata_InteractionType = 4 // Professionally produced audio (eg. TV Show, Podcast). RecognitionMetadata_PROFESSIONALLY_PRODUCED RecognitionMetadata_InteractionType = 5 // Transcribe spoken questions and queries into text. RecognitionMetadata_VOICE_SEARCH RecognitionMetadata_InteractionType = 6 // Transcribe voice commands, such as for controlling a device. RecognitionMetadata_VOICE_COMMAND RecognitionMetadata_InteractionType = 7 // Transcribe speech to text to create a written document, such as a // text-message, email or report. RecognitionMetadata_DICTATION RecognitionMetadata_InteractionType = 8 )
func (RecognitionMetadata_InteractionType) EnumDescriptor ¶
func (RecognitionMetadata_InteractionType) EnumDescriptor() ([]byte, []int)
func (RecognitionMetadata_InteractionType) String ¶
func (x RecognitionMetadata_InteractionType) String() string
type RecognitionMetadata_MicrophoneDistance ¶
type RecognitionMetadata_MicrophoneDistance int32
Enumerates the types of capture settings describing an audio file.
const ( // Audio type is not known. RecognitionMetadata_MICROPHONE_DISTANCE_UNSPECIFIED RecognitionMetadata_MicrophoneDistance = 0 // The audio was captured from a closely placed microphone. Eg. phone, // dictaphone, or handheld microphone. Generally if there speaker is within // 1 meter of the microphone. RecognitionMetadata_NEARFIELD RecognitionMetadata_MicrophoneDistance = 1 // The speaker if within 3 meters of the microphone. RecognitionMetadata_MIDFIELD RecognitionMetadata_MicrophoneDistance = 2 // The speaker is more than 3 meters away from the microphone. RecognitionMetadata_FARFIELD RecognitionMetadata_MicrophoneDistance = 3 )
func (RecognitionMetadata_MicrophoneDistance) EnumDescriptor ¶
func (RecognitionMetadata_MicrophoneDistance) EnumDescriptor() ([]byte, []int)
func (RecognitionMetadata_MicrophoneDistance) String ¶
func (x RecognitionMetadata_MicrophoneDistance) String() string
type RecognitionMetadata_OriginalMediaType ¶
type RecognitionMetadata_OriginalMediaType int32
The original media the speech was recorded on.
const ( // Unknown original media type. RecognitionMetadata_ORIGINAL_MEDIA_TYPE_UNSPECIFIED RecognitionMetadata_OriginalMediaType = 0 // The speech data is an audio recording. RecognitionMetadata_AUDIO RecognitionMetadata_OriginalMediaType = 1 // The speech data originally recorded on a video. RecognitionMetadata_VIDEO RecognitionMetadata_OriginalMediaType = 2 )
func (RecognitionMetadata_OriginalMediaType) EnumDescriptor ¶
func (RecognitionMetadata_OriginalMediaType) EnumDescriptor() ([]byte, []int)
func (RecognitionMetadata_OriginalMediaType) String ¶
func (x RecognitionMetadata_OriginalMediaType) String() string
type RecognitionMetadata_RecordingDeviceType ¶
type RecognitionMetadata_RecordingDeviceType int32
The type of device the speech was recorded with.
const ( // The recording device is unknown. RecognitionMetadata_RECORDING_DEVICE_TYPE_UNSPECIFIED RecognitionMetadata_RecordingDeviceType = 0 // Speech was recorded on a smartphone. RecognitionMetadata_SMARTPHONE RecognitionMetadata_RecordingDeviceType = 1 // Speech was recorded using a personal computer or tablet. RecognitionMetadata_PC RecognitionMetadata_RecordingDeviceType = 2 // Speech was recorded over a phone line. RecognitionMetadata_PHONE_LINE RecognitionMetadata_RecordingDeviceType = 3 // Speech was recorded in a vehicle. RecognitionMetadata_VEHICLE RecognitionMetadata_RecordingDeviceType = 4 // Speech was recorded outdoors. RecognitionMetadata_OTHER_OUTDOOR_DEVICE RecognitionMetadata_RecordingDeviceType = 5 // Speech was recorded indoors. RecognitionMetadata_OTHER_INDOOR_DEVICE RecognitionMetadata_RecordingDeviceType = 6 )
func (RecognitionMetadata_RecordingDeviceType) EnumDescriptor ¶
func (RecognitionMetadata_RecordingDeviceType) EnumDescriptor() ([]byte, []int)
func (RecognitionMetadata_RecordingDeviceType) String ¶
func (x RecognitionMetadata_RecordingDeviceType) String() string
type RecognizeRequest ¶
type RecognizeRequest struct { // *Required* Provides information to the recognizer that specifies how to // process the request. Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config" json:"config,omitempty"` // *Required* The audio data to be recognized. Audio *RecognitionAudio `protobuf:"bytes,2,opt,name=audio" json:"audio,omitempty"` }
The top-level message sent by the client for the `Recognize` method.
func (*RecognizeRequest) Descriptor ¶
func (*RecognizeRequest) Descriptor() ([]byte, []int)
func (*RecognizeRequest) GetAudio ¶
func (m *RecognizeRequest) GetAudio() *RecognitionAudio
func (*RecognizeRequest) GetConfig ¶
func (m *RecognizeRequest) GetConfig() *RecognitionConfig
func (*RecognizeRequest) ProtoMessage ¶
func (*RecognizeRequest) ProtoMessage()
func (*RecognizeRequest) Reset ¶
func (m *RecognizeRequest) Reset()
func (*RecognizeRequest) String ¶
func (m *RecognizeRequest) String() string
type RecognizeResponse ¶
type RecognizeResponse struct { // Output only. Sequential list of transcription results corresponding to // sequential portions of audio. Results []*SpeechRecognitionResult `protobuf:"bytes,2,rep,name=results" json:"results,omitempty"` }
The only message returned to the client by the `Recognize` method. It contains the result as zero or more sequential `SpeechRecognitionResult` messages.
func (*RecognizeResponse) Descriptor ¶
func (*RecognizeResponse) Descriptor() ([]byte, []int)
func (*RecognizeResponse) GetResults ¶
func (m *RecognizeResponse) GetResults() []*SpeechRecognitionResult
func (*RecognizeResponse) ProtoMessage ¶
func (*RecognizeResponse) ProtoMessage()
func (*RecognizeResponse) Reset ¶
func (m *RecognizeResponse) Reset()
func (*RecognizeResponse) String ¶
func (m *RecognizeResponse) String() string
type SpeechClient ¶
type SpeechClient interface { // Performs synchronous speech recognition: receive results after all audio // has been sent and processed. Recognize(ctx context.Context, in *RecognizeRequest, opts ...grpc.CallOption) (*RecognizeResponse, error) // Performs asynchronous speech recognition: receive results via the // google.longrunning.Operations interface. Returns either an // `Operation.error` or an `Operation.response` which contains // a `LongRunningRecognizeResponse` message. LongRunningRecognize(ctx context.Context, in *LongRunningRecognizeRequest, opts ...grpc.CallOption) (*google_longrunning.Operation, error) // Performs bidirectional streaming speech recognition: receive results while // sending audio. This method is only available via the gRPC API (not REST). StreamingRecognize(ctx context.Context, opts ...grpc.CallOption) (Speech_StreamingRecognizeClient, error) }
func NewSpeechClient ¶
func NewSpeechClient(cc *grpc.ClientConn) SpeechClient
type SpeechContext ¶
type SpeechContext struct { // *Optional* A list of strings containing words and phrases "hints" so that // the speech recognition is more likely to recognize them. This can be used // to improve the accuracy for specific words and phrases, for example, if // specific commands are typically spoken by the user. This can also be used // to add additional words to the vocabulary of the recognizer. See // [usage limits](https://cloud.google.com/speech/limits#content). Phrases []string `protobuf:"bytes,1,rep,name=phrases" json:"phrases,omitempty"` }
Provides "hints" to the speech recognizer to favor specific words and phrases in the results.
func (*SpeechContext) Descriptor ¶
func (*SpeechContext) Descriptor() ([]byte, []int)
func (*SpeechContext) GetPhrases ¶
func (m *SpeechContext) GetPhrases() []string
func (*SpeechContext) ProtoMessage ¶
func (*SpeechContext) ProtoMessage()
func (*SpeechContext) Reset ¶
func (m *SpeechContext) Reset()
func (*SpeechContext) String ¶
func (m *SpeechContext) String() string
type SpeechRecognitionAlternative ¶
type SpeechRecognitionAlternative struct { // Output only. Transcript text representing the words that the user spoke. Transcript string `protobuf:"bytes,1,opt,name=transcript" json:"transcript,omitempty"` // Output only. The confidence estimate between 0.0 and 1.0. A higher number // indicates an estimated greater likelihood that the recognized words are // correct. This field is set only for the top alternative of a non-streaming // result or, of a streaming result where `is_final=true`. // This field is not guaranteed to be accurate and users should not rely on it // to be always provided. // The default of 0.0 is a sentinel value indicating `confidence` was not set. Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"` // Output only. A list of word-specific information for each recognized word. // Note: When enable_speaker_diarization is true, you will see all the words // from the beginning of the audio. Words []*WordInfo `protobuf:"bytes,3,rep,name=words" json:"words,omitempty"` }
Alternative hypotheses (a.k.a. n-best list).
func (*SpeechRecognitionAlternative) Descriptor ¶
func (*SpeechRecognitionAlternative) Descriptor() ([]byte, []int)
func (*SpeechRecognitionAlternative) GetConfidence ¶
func (m *SpeechRecognitionAlternative) GetConfidence() float32
func (*SpeechRecognitionAlternative) GetTranscript ¶
func (m *SpeechRecognitionAlternative) GetTranscript() string
func (*SpeechRecognitionAlternative) GetWords ¶
func (m *SpeechRecognitionAlternative) GetWords() []*WordInfo
func (*SpeechRecognitionAlternative) ProtoMessage ¶
func (*SpeechRecognitionAlternative) ProtoMessage()
func (*SpeechRecognitionAlternative) Reset ¶
func (m *SpeechRecognitionAlternative) Reset()
func (*SpeechRecognitionAlternative) String ¶
func (m *SpeechRecognitionAlternative) String() string
type SpeechRecognitionResult ¶
type SpeechRecognitionResult struct { // Output only. May contain one or more recognition hypotheses (up to the // maximum specified in `max_alternatives`). // These alternatives are ordered in terms of accuracy, with the top (first) // alternative being the most probable, as ranked by the recognizer. Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives" json:"alternatives,omitempty"` }
A speech recognition result corresponding to a portion of the audio.
func (*SpeechRecognitionResult) Descriptor ¶
func (*SpeechRecognitionResult) Descriptor() ([]byte, []int)
func (*SpeechRecognitionResult) GetAlternatives ¶
func (m *SpeechRecognitionResult) GetAlternatives() []*SpeechRecognitionAlternative
func (*SpeechRecognitionResult) ProtoMessage ¶
func (*SpeechRecognitionResult) ProtoMessage()
func (*SpeechRecognitionResult) Reset ¶
func (m *SpeechRecognitionResult) Reset()
func (*SpeechRecognitionResult) String ¶
func (m *SpeechRecognitionResult) String() string
type SpeechServer ¶
type SpeechServer interface { // Performs synchronous speech recognition: receive results after all audio // has been sent and processed. Recognize(context.Context, *RecognizeRequest) (*RecognizeResponse, error) // Performs asynchronous speech recognition: receive results via the // google.longrunning.Operations interface. Returns either an // `Operation.error` or an `Operation.response` which contains // a `LongRunningRecognizeResponse` message. LongRunningRecognize(context.Context, *LongRunningRecognizeRequest) (*google_longrunning.Operation, error) // Performs bidirectional streaming speech recognition: receive results while // sending audio. This method is only available via the gRPC API (not REST). StreamingRecognize(Speech_StreamingRecognizeServer) error }
type Speech_StreamingRecognizeClient ¶
type Speech_StreamingRecognizeClient interface { Send(*StreamingRecognizeRequest) error Recv() (*StreamingRecognizeResponse, error) grpc.ClientStream }
type Speech_StreamingRecognizeServer ¶
type Speech_StreamingRecognizeServer interface { Send(*StreamingRecognizeResponse) error Recv() (*StreamingRecognizeRequest, error) grpc.ServerStream }
type StreamingRecognitionConfig ¶
type StreamingRecognitionConfig struct { // *Required* Provides information to the recognizer that specifies how to // process the request. Config *RecognitionConfig `protobuf:"bytes,1,opt,name=config" json:"config,omitempty"` // *Optional* If `false` or omitted, the recognizer will perform continuous // recognition (continuing to wait for and process audio even if the user // pauses speaking) until the client closes the input stream (gRPC API) or // until the maximum time limit has been reached. May return multiple // `StreamingRecognitionResult`s with the `is_final` flag set to `true`. // // If `true`, the recognizer will detect a single spoken utterance. When it // detects that the user has paused or stopped speaking, it will return an // `END_OF_SINGLE_UTTERANCE` event and cease recognition. It will return no // more than one `StreamingRecognitionResult` with the `is_final` flag set to // `true`. SingleUtterance bool `protobuf:"varint,2,opt,name=single_utterance,json=singleUtterance" json:"single_utterance,omitempty"` // *Optional* If `true`, interim results (tentative hypotheses) may be // returned as they become available (these interim results are indicated with // the `is_final=false` flag). // If `false` or omitted, only `is_final=true` result(s) are returned. InterimResults bool `protobuf:"varint,3,opt,name=interim_results,json=interimResults" json:"interim_results,omitempty"` }
Provides information to the recognizer that specifies how to process the request.
func (*StreamingRecognitionConfig) Descriptor ¶
func (*StreamingRecognitionConfig) Descriptor() ([]byte, []int)
func (*StreamingRecognitionConfig) GetConfig ¶
func (m *StreamingRecognitionConfig) GetConfig() *RecognitionConfig
func (*StreamingRecognitionConfig) GetInterimResults ¶
func (m *StreamingRecognitionConfig) GetInterimResults() bool
func (*StreamingRecognitionConfig) GetSingleUtterance ¶
func (m *StreamingRecognitionConfig) GetSingleUtterance() bool
func (*StreamingRecognitionConfig) ProtoMessage ¶
func (*StreamingRecognitionConfig) ProtoMessage()
func (*StreamingRecognitionConfig) Reset ¶
func (m *StreamingRecognitionConfig) Reset()
func (*StreamingRecognitionConfig) String ¶
func (m *StreamingRecognitionConfig) String() string
type StreamingRecognitionResult ¶
type StreamingRecognitionResult struct { // Output only. May contain one or more recognition hypotheses (up to the // maximum specified in `max_alternatives`). // These alternatives are ordered in terms of accuracy, with the top (first) // alternative being the most probable, as ranked by the recognizer. Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives" json:"alternatives,omitempty"` // Output only. If `false`, this `StreamingRecognitionResult` represents an // interim result that may change. If `true`, this is the final time the // speech service will return this particular `StreamingRecognitionResult`, // the recognizer will not return any further hypotheses for this portion of // the transcript and corresponding audio. IsFinal bool `protobuf:"varint,2,opt,name=is_final,json=isFinal" json:"is_final,omitempty"` // Output only. An estimate of the likelihood that the recognizer will not // change its guess about this interim result. Values range from 0.0 // (completely unstable) to 1.0 (completely stable). // This field is only provided for interim results (`is_final=false`). // The default of 0.0 is a sentinel value indicating `stability` was not set. Stability float32 `protobuf:"fixed32,3,opt,name=stability" json:"stability,omitempty"` }
A streaming speech recognition result corresponding to a portion of the audio that is currently being processed.
func (*StreamingRecognitionResult) Descriptor ¶
func (*StreamingRecognitionResult) Descriptor() ([]byte, []int)
func (*StreamingRecognitionResult) GetAlternatives ¶
func (m *StreamingRecognitionResult) GetAlternatives() []*SpeechRecognitionAlternative
func (*StreamingRecognitionResult) GetIsFinal ¶
func (m *StreamingRecognitionResult) GetIsFinal() bool
func (*StreamingRecognitionResult) GetStability ¶
func (m *StreamingRecognitionResult) GetStability() float32
func (*StreamingRecognitionResult) ProtoMessage ¶
func (*StreamingRecognitionResult) ProtoMessage()
func (*StreamingRecognitionResult) Reset ¶
func (m *StreamingRecognitionResult) Reset()
func (*StreamingRecognitionResult) String ¶
func (m *StreamingRecognitionResult) String() string
type StreamingRecognizeRequest ¶
type StreamingRecognizeRequest struct { // The streaming request, which is either a streaming config or audio content. // // Types that are valid to be assigned to StreamingRequest: // *StreamingRecognizeRequest_StreamingConfig // *StreamingRecognizeRequest_AudioContent StreamingRequest isStreamingRecognizeRequest_StreamingRequest `protobuf_oneof:"streaming_request"` }
The top-level message sent by the client for the `StreamingRecognize` method. Multiple `StreamingRecognizeRequest` messages are sent. The first message must contain a `streaming_config` message and must not contain `audio` data. All subsequent messages must contain `audio` data and must not contain a `streaming_config` message.
func (*StreamingRecognizeRequest) Descriptor ¶
func (*StreamingRecognizeRequest) Descriptor() ([]byte, []int)
func (*StreamingRecognizeRequest) GetAudioContent ¶
func (m *StreamingRecognizeRequest) GetAudioContent() []byte
func (*StreamingRecognizeRequest) GetStreamingConfig ¶
func (m *StreamingRecognizeRequest) GetStreamingConfig() *StreamingRecognitionConfig
func (*StreamingRecognizeRequest) GetStreamingRequest ¶
func (m *StreamingRecognizeRequest) GetStreamingRequest() isStreamingRecognizeRequest_StreamingRequest
func (*StreamingRecognizeRequest) ProtoMessage ¶
func (*StreamingRecognizeRequest) ProtoMessage()
func (*StreamingRecognizeRequest) Reset ¶
func (m *StreamingRecognizeRequest) Reset()
func (*StreamingRecognizeRequest) String ¶
func (m *StreamingRecognizeRequest) String() string
func (*StreamingRecognizeRequest) XXX_OneofFuncs ¶
func (*StreamingRecognizeRequest) XXX_OneofFuncs() (func(msg proto.Message, b *proto.Buffer) error, func(msg proto.Message, tag, wire int, b *proto.Buffer) (bool, error), func(msg proto.Message) (n int), []interface{})
XXX_OneofFuncs is for the internal use of the proto package.
type StreamingRecognizeRequest_AudioContent ¶
type StreamingRecognizeRequest_AudioContent struct {
AudioContent []byte `protobuf:"bytes,2,opt,name=audio_content,json=audioContent,proto3,oneof"`
}
type StreamingRecognizeRequest_StreamingConfig ¶
type StreamingRecognizeRequest_StreamingConfig struct {
StreamingConfig *StreamingRecognitionConfig `protobuf:"bytes,1,opt,name=streaming_config,json=streamingConfig,oneof"`
}
type StreamingRecognizeResponse ¶
type StreamingRecognizeResponse struct { // Output only. If set, returns a [google.rpc.Status][google.rpc.Status] message that // specifies the error for the operation. Error *google_rpc.Status `protobuf:"bytes,1,opt,name=error" json:"error,omitempty"` // Output only. This repeated list contains zero or more results that // correspond to consecutive portions of the audio currently being processed. // It contains zero or one `is_final=true` result (the newly settled portion), // followed by zero or more `is_final=false` results (the interim results). Results []*StreamingRecognitionResult `protobuf:"bytes,2,rep,name=results" json:"results,omitempty"` // Output only. Indicates the type of speech event. SpeechEventType StreamingRecognizeResponse_SpeechEventType `` /* 180-byte string literal not displayed */ }
`StreamingRecognizeResponse` is the only message returned to the client by `StreamingRecognize`. A series of zero or more `StreamingRecognizeResponse` messages are streamed back to the client. If there is no recognizable audio, and `single_utterance` is set to false, then no messages are streamed back to the client.
Here's an example of a series of ten `StreamingRecognizeResponse`s that might be returned while processing audio:
1. results { alternatives { transcript: "tube" } stability: 0.01 }
2. results { alternatives { transcript: "to be a" } stability: 0.01 }
- results { alternatives { transcript: "to be" } stability: 0.9 } results { alternatives { transcript: " or not to be" } stability: 0.01 }
- results { alternatives { transcript: "to be or not to be" confidence: 0.92 } alternatives { transcript: "to bee or not to bee" } is_final: true }
5. results { alternatives { transcript: " that's" } stability: 0.01 }
- results { alternatives { transcript: " that is" } stability: 0.9 } results { alternatives { transcript: " the question" } stability: 0.01 }
- results { alternatives { transcript: " that is the question" confidence: 0.98 } alternatives { transcript: " that was the question" } is_final: true }
Notes:
- Only two of the above responses #4 and #7 contain final results; they are indicated by `is_final: true`. Concatenating these together generates the full transcript: "to be or not to be that is the question".
- The others contain interim `results`. #3 and #6 contain two interim `results`: the first portion has a high stability and is less likely to change; the second portion has a low stability and is very likely to change. A UI designer might choose to show only high stability `results`.
- The specific `stability` and `confidence` values shown above are only for illustrative purposes. Actual values may vary.
- In each response, only one of these fields will be set: `error`, `speech_event_type`, or one or more (repeated) `results`.
func (*StreamingRecognizeResponse) Descriptor ¶
func (*StreamingRecognizeResponse) Descriptor() ([]byte, []int)
func (*StreamingRecognizeResponse) GetError ¶
func (m *StreamingRecognizeResponse) GetError() *google_rpc.Status
func (*StreamingRecognizeResponse) GetResults ¶
func (m *StreamingRecognizeResponse) GetResults() []*StreamingRecognitionResult
func (*StreamingRecognizeResponse) GetSpeechEventType ¶
func (m *StreamingRecognizeResponse) GetSpeechEventType() StreamingRecognizeResponse_SpeechEventType
func (*StreamingRecognizeResponse) ProtoMessage ¶
func (*StreamingRecognizeResponse) ProtoMessage()
func (*StreamingRecognizeResponse) Reset ¶
func (m *StreamingRecognizeResponse) Reset()
func (*StreamingRecognizeResponse) String ¶
func (m *StreamingRecognizeResponse) String() string
type StreamingRecognizeResponse_SpeechEventType ¶
type StreamingRecognizeResponse_SpeechEventType int32
Indicates the type of speech event.
const ( // No speech event specified. StreamingRecognizeResponse_SPEECH_EVENT_UNSPECIFIED StreamingRecognizeResponse_SpeechEventType = 0 // This event indicates that the server has detected the end of the user's // speech utterance and expects no additional speech. Therefore, the server // will not process additional audio (although it may subsequently return // additional results). The client should stop sending additional audio // data, half-close the gRPC connection, and wait for any additional results // until the server closes the gRPC connection. This event is only sent if // `single_utterance` was set to `true`, and is not used otherwise. StreamingRecognizeResponse_END_OF_SINGLE_UTTERANCE StreamingRecognizeResponse_SpeechEventType = 1 )
func (StreamingRecognizeResponse_SpeechEventType) EnumDescriptor ¶
func (StreamingRecognizeResponse_SpeechEventType) EnumDescriptor() ([]byte, []int)
func (StreamingRecognizeResponse_SpeechEventType) String ¶
func (x StreamingRecognizeResponse_SpeechEventType) String() string
type WordInfo ¶
type WordInfo struct { // Output only. Time offset relative to the beginning of the audio, // and corresponding to the start of the spoken word. // This field is only set if `enable_word_time_offsets=true` and only // in the top hypothesis. // This is an experimental feature and the accuracy of the time offset can // vary. StartTime *google_protobuf3.Duration `protobuf:"bytes,1,opt,name=start_time,json=startTime" json:"start_time,omitempty"` // Output only. Time offset relative to the beginning of the audio, // and corresponding to the end of the spoken word. // This field is only set if `enable_word_time_offsets=true` and only // in the top hypothesis. // This is an experimental feature and the accuracy of the time offset can // vary. EndTime *google_protobuf3.Duration `protobuf:"bytes,2,opt,name=end_time,json=endTime" json:"end_time,omitempty"` // Output only. The word corresponding to this set of information. Word string `protobuf:"bytes,3,opt,name=word" json:"word,omitempty"` }
Word-specific information for recognized words.
func (*WordInfo) Descriptor ¶
func (*WordInfo) GetEndTime ¶
func (m *WordInfo) GetEndTime() *google_protobuf3.Duration
func (*WordInfo) GetStartTime ¶
func (m *WordInfo) GetStartTime() *google_protobuf3.Duration
func (*WordInfo) ProtoMessage ¶
func (*WordInfo) ProtoMessage()