Reference documentation and code samples for the Cloud Speech V1p1beta1 Client class StreamingRecognitionConfig.
Provides information to the recognizer that specifies how to process the request.
Generated from protobuf message google.cloud.speech.v1p1beta1.StreamingRecognitionConfig
Namespace
Google \ Cloud \ Speech \ V1p1beta1Methods
__construct
Constructor.
data
array
Optional. Data for populating the Message object.
↳ config
RecognitionConfig
Required. Provides information to the recognizer that specifies how to process the request.
↳ single_utterance
bool
If false
or omitted, the recognizer will perform continuous recognition (continuing to wait for and process audio even if the user pauses speaking) until the client closes the input stream (gRPC API) or until the maximum time limit has been reached. May return multiple StreamingRecognitionResult
s with the is_final
flag set to true
. If true
, the recognizer will detect a single spoken utterance. When it detects that the user has paused or stopped speaking, it will return an END_OF_SINGLE_UTTERANCE
event and cease recognition. It will return no more than one StreamingRecognitionResult
with the is_final
flag set to true
. The single_utterance
field can only be used with specified models, otherwise an error is thrown. The model
field in RecognitionConfig
must be set to: * * command_and_search
* * phone_call
AND additional field useEnhanced
= true
* * The model
field is left undefined. In this case the API auto-selects a model based on any other parameters that you set in RecognitionConfig
.
↳ interim_results
bool
If true
, interim results (tentative hypotheses) may be returned as they become available (these interim results are indicated with the is_final=false
flag). If false
or omitted, only is_final=true
result(s) are returned.
↳ enable_voice_activity_events
bool
If true
, responses with voice activity speech events will be returned as they are detected.
↳ voice_activity_timeout
StreamingRecognitionConfig\VoiceActivityTimeout
If set, the server will automatically close the stream after the specified duration has elapsed after the last VOICE_ACTIVITY speech event has been sent. The field voice_activity_events
must also be set to true.
getConfig
Required. Provides information to the recognizer that specifies how to process the request.
hasConfig
clearConfig
setConfig
Required. Provides information to the recognizer that specifies how to process the request.
$this
getSingleUtterance
If false
or omitted, the recognizer will perform continuous
recognition (continuing to wait for and process audio even if the user
pauses speaking) until the client closes the input stream (gRPC API) or
until the maximum time limit has been reached. May return multiple StreamingRecognitionResult
s with the is_final
flag set to true
.
If true
, the recognizer will detect a single spoken utterance. When it
detects that the user has paused or stopped speaking, it will return an END_OF_SINGLE_UTTERANCE
event and cease recognition. It will return no
more than one StreamingRecognitionResult
with the is_final
flag set to true
.
The single_utterance
field can only be used with specified models,
otherwise an error is thrown. The model
field in RecognitionConfig
must be set to:
-
command_and_search
-
phone_call
AND additional fielduseEnhanced
=true
- The
model
field is left undefined. In this case the API auto-selects a model based on any other parameters that you set inRecognitionConfig
.
bool
setSingleUtterance
If false
or omitted, the recognizer will perform continuous
recognition (continuing to wait for and process audio even if the user
pauses speaking) until the client closes the input stream (gRPC API) or
until the maximum time limit has been reached. May return multiple StreamingRecognitionResult
s with the is_final
flag set to true
.
If true
, the recognizer will detect a single spoken utterance. When it
detects that the user has paused or stopped speaking, it will return an END_OF_SINGLE_UTTERANCE
event and cease recognition. It will return no
more than one StreamingRecognitionResult
with the is_final
flag set to true
.
The single_utterance
field can only be used with specified models,
otherwise an error is thrown. The model
field in RecognitionConfig
must be set to:
-
command_and_search
-
phone_call
AND additional fielduseEnhanced
=true
- The
model
field is left undefined. In this case the API auto-selects a model based on any other parameters that you set inRecognitionConfig
.
var
bool
$this
getInterimResults
If true
, interim results (tentative hypotheses) may be
returned as they become available (these interim results are indicated with
the is_final=false
flag).
If false
or omitted, only is_final=true
result(s) are returned.
bool
setInterimResults
If true
, interim results (tentative hypotheses) may be
returned as they become available (these interim results are indicated with
the is_final=false
flag).
If false
or omitted, only is_final=true
result(s) are returned.
var
bool
$this
getEnableVoiceActivityEvents
If true
, responses with voice activity speech events will be returned as
they are detected.
bool
setEnableVoiceActivityEvents
If true
, responses with voice activity speech events will be returned as
they are detected.
var
bool
$this
getVoiceActivityTimeout
If set, the server will automatically close the stream after the specified
duration has elapsed after the last VOICE_ACTIVITY speech event has been
sent. The field voice_activity_events
must also be set to true.
hasVoiceActivityTimeout
clearVoiceActivityTimeout
setVoiceActivityTimeout
If set, the server will automatically close the stream after the specified
duration has elapsed after the last VOICE_ACTIVITY speech event has been
sent. The field voice_activity_events
must also be set to true.
$this