Package com.google.cloud.speech.v1beta1
Interface StreamingRecognitionConfigOrBuilder
-
- All Superinterfaces:
com.google.protobuf.MessageLiteOrBuilder
,com.google.protobuf.MessageOrBuilder
- All Known Implementing Classes:
StreamingRecognitionConfig
,StreamingRecognitionConfig.Builder
public interface StreamingRecognitionConfigOrBuilder extends com.google.protobuf.MessageOrBuilder
-
-
Method Summary
All Methods Instance Methods Abstract Methods Modifier and Type Method Description RecognitionConfig
getConfig()
*Required* Provides information to the recognizer that specifies how to process the request.RecognitionConfigOrBuilder
getConfigOrBuilder()
*Required* Provides information to the recognizer that specifies how to process the request.boolean
getInterimResults()
*Optional* If `true`, interim results (tentative hypotheses) may be returned as they become available (these interim results are indicated with the `is_final=false` flag).boolean
getSingleUtterance()
*Optional* If `false` or omitted, the recognizer will perform continuous recognition (continuing to wait for and process audio even if the user pauses speaking) until the client closes the input stream (gRPC API) or until the maximum time limit has been reached.boolean
hasConfig()
*Required* Provides information to the recognizer that specifies how to process the request.-
Methods inherited from interface com.google.protobuf.MessageOrBuilder
findInitializationErrors, getAllFields, getDefaultInstanceForType, getDescriptorForType, getField, getInitializationErrorString, getOneofFieldDescriptor, getRepeatedField, getRepeatedFieldCount, getUnknownFields, hasField, hasOneof
-
-
-
-
Method Detail
-
hasConfig
boolean hasConfig()
*Required* Provides information to the recognizer that specifies how to process the request.
.google.cloud.speech.v1beta1.RecognitionConfig config = 1;
-
getConfig
RecognitionConfig getConfig()
*Required* Provides information to the recognizer that specifies how to process the request.
.google.cloud.speech.v1beta1.RecognitionConfig config = 1;
-
getConfigOrBuilder
RecognitionConfigOrBuilder getConfigOrBuilder()
*Required* Provides information to the recognizer that specifies how to process the request.
.google.cloud.speech.v1beta1.RecognitionConfig config = 1;
-
getSingleUtterance
boolean getSingleUtterance()
*Optional* If `false` or omitted, the recognizer will perform continuous recognition (continuing to wait for and process audio even if the user pauses speaking) until the client closes the input stream (gRPC API) or until the maximum time limit has been reached. May return multiple `StreamingRecognitionResult`s with the `is_final` flag set to `true`. If `true`, the recognizer will detect a single spoken utterance. When it detects that the user has paused or stopped speaking, it will return an `END_OF_UTTERANCE` event and cease recognition. It will return no more than one `StreamingRecognitionResult` with the `is_final` flag set to `true`.
bool single_utterance = 2;
-
getInterimResults
boolean getInterimResults()
*Optional* If `true`, interim results (tentative hypotheses) may be returned as they become available (these interim results are indicated with the `is_final=false` flag). If `false` or omitted, only `is_final=true` result(s) are returned.
bool interim_results = 3;
-
-