Interface StreamingRecognitionConfigOrBuilder

    • Method Summary

      All Methods Instance Methods Abstract Methods 
      Modifier and Type Method Description
      RecognitionConfig getConfig()
      Required.
      RecognitionConfigOrBuilder getConfigOrBuilder()
      Required.
      boolean getEnableVoiceActivityEvents()
      If `true`, responses with voice activity speech events will be returned as they are detected.
      boolean getInterimResults()
      If `true`, interim results (tentative hypotheses) may be returned as they become available (these interim results are indicated with the `is_final=false` flag).
      boolean getSingleUtterance()
      If `false` or omitted, the recognizer will perform continuous recognition (continuing to wait for and process audio even if the user pauses speaking) until the client closes the input stream (gRPC API) or until the maximum time limit has been reached.
      StreamingRecognitionConfig.VoiceActivityTimeout getVoiceActivityTimeout()
      If set, the server will automatically close the stream after the specified duration has elapsed after the last VOICE_ACTIVITY speech event has been sent.
      StreamingRecognitionConfig.VoiceActivityTimeoutOrBuilder getVoiceActivityTimeoutOrBuilder()
      If set, the server will automatically close the stream after the specified duration has elapsed after the last VOICE_ACTIVITY speech event has been sent.
      boolean hasConfig()
      Required.
      boolean hasVoiceActivityTimeout()
      If set, the server will automatically close the stream after the specified duration has elapsed after the last VOICE_ACTIVITY speech event has been sent.
      • Methods inherited from interface com.google.protobuf.MessageLiteOrBuilder

        isInitialized
      • Methods inherited from interface com.google.protobuf.MessageOrBuilder

        findInitializationErrors, getAllFields, getDefaultInstanceForType, getDescriptorForType, getField, getInitializationErrorString, getOneofFieldDescriptor, getRepeatedField, getRepeatedFieldCount, getUnknownFields, hasField, hasOneof
    • Method Detail

      • hasConfig

        boolean hasConfig()
         Required. Provides information to the recognizer that specifies how to
         process the request.
         
        .google.cloud.speech.v1.RecognitionConfig config = 1 [(.google.api.field_behavior) = REQUIRED];
        Returns:
        Whether the config field is set.
      • getConfig

        RecognitionConfig getConfig()
         Required. Provides information to the recognizer that specifies how to
         process the request.
         
        .google.cloud.speech.v1.RecognitionConfig config = 1 [(.google.api.field_behavior) = REQUIRED];
        Returns:
        The config.
      • getConfigOrBuilder

        RecognitionConfigOrBuilder getConfigOrBuilder()
         Required. Provides information to the recognizer that specifies how to
         process the request.
         
        .google.cloud.speech.v1.RecognitionConfig config = 1 [(.google.api.field_behavior) = REQUIRED];
      • getSingleUtterance

        boolean getSingleUtterance()
         If `false` or omitted, the recognizer will perform continuous
         recognition (continuing to wait for and process audio even if the user
         pauses speaking) until the client closes the input stream (gRPC API) or
         until the maximum time limit has been reached. May return multiple
         `StreamingRecognitionResult`s with the `is_final` flag set to `true`.
        
         If `true`, the recognizer will detect a single spoken utterance. When it
         detects that the user has paused or stopped speaking, it will return an
         `END_OF_SINGLE_UTTERANCE` event and cease recognition. It will return no
         more than one `StreamingRecognitionResult` with the `is_final` flag set to
         `true`.
        
         The `single_utterance` field can only be used with specified models,
         otherwise an error is thrown. The `model` field in [`RecognitionConfig`][]
         must be set to:
        
         * `command_and_search`
         * `phone_call` AND additional field `useEnhanced`=`true`
         * The `model` field is left undefined. In this case the API auto-selects
           a model based on any other parameters that you set in
           `RecognitionConfig`.
         
        bool single_utterance = 2;
        Returns:
        The singleUtterance.
      • getInterimResults

        boolean getInterimResults()
         If `true`, interim results (tentative hypotheses) may be
         returned as they become available (these interim results are indicated with
         the `is_final=false` flag).
         If `false` or omitted, only `is_final=true` result(s) are returned.
         
        bool interim_results = 3;
        Returns:
        The interimResults.
      • getEnableVoiceActivityEvents

        boolean getEnableVoiceActivityEvents()
         If `true`, responses with voice activity speech events will be returned as
         they are detected.
         
        bool enable_voice_activity_events = 5;
        Returns:
        The enableVoiceActivityEvents.
      • hasVoiceActivityTimeout

        boolean hasVoiceActivityTimeout()
         If set, the server will automatically close the stream after the specified
         duration has elapsed after the last VOICE_ACTIVITY speech event has been
         sent. The field `voice_activity_events` must also be set to true.
         
        .google.cloud.speech.v1.StreamingRecognitionConfig.VoiceActivityTimeout voice_activity_timeout = 6;
        Returns:
        Whether the voiceActivityTimeout field is set.
      • getVoiceActivityTimeout

        StreamingRecognitionConfig.VoiceActivityTimeout getVoiceActivityTimeout()
         If set, the server will automatically close the stream after the specified
         duration has elapsed after the last VOICE_ACTIVITY speech event has been
         sent. The field `voice_activity_events` must also be set to true.
         
        .google.cloud.speech.v1.StreamingRecognitionConfig.VoiceActivityTimeout voice_activity_timeout = 6;
        Returns:
        The voiceActivityTimeout.
      • getVoiceActivityTimeoutOrBuilder

        StreamingRecognitionConfig.VoiceActivityTimeoutOrBuilder getVoiceActivityTimeoutOrBuilder()
         If set, the server will automatically close the stream after the specified
         duration has elapsed after the last VOICE_ACTIVITY speech event has been
         sent. The field `voice_activity_events` must also be set to true.
         
        .google.cloud.speech.v1.StreamingRecognitionConfig.VoiceActivityTimeout voice_activity_timeout = 6;