diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionAudio.java b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionAudio.java index cf2b2ac1e08c..e69a6267355d 100644 --- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionAudio.java +++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionAudio.java @@ -8,7 +8,7 @@ * Contains audio data in the encoding specified in the `RecognitionConfig`. * Either `content` or `uri` must be supplied. Supplying both or neither * returns [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]. See - * [audio limits](https://cloud.google.com/speech/limits#content). + * [content limits](/speech-to-text/quotas#content). * * * Protobuf type {@code google.cloud.speech.v1p1beta1.RecognitionAudio} @@ -151,7 +151,8 @@ public com.google.protobuf.ByteString getContent() { /** *
* URI that points to a file that contains audio data bytes as specified in
- * `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
+ * `RecognitionConfig`. The file must not be compressed (for example, gzip).
+ * Currently, only Google Cloud Storage URIs are
* supported, which must be specified in the following format:
* `gs://bucket_name/object_name` (other URI formats return
* [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
@@ -180,7 +181,8 @@ public java.lang.String getUri() {
/**
*
* URI that points to a file that contains audio data bytes as specified in
- * `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
+ * `RecognitionConfig`. The file must not be compressed (for example, gzip).
+ * Currently, only Google Cloud Storage URIs are
* supported, which must be specified in the following format:
* `gs://bucket_name/object_name` (other URI formats return
* [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
@@ -400,7 +402,7 @@ protected Builder newBuilderForType(
* Contains audio data in the encoding specified in the `RecognitionConfig`.
* Either `content` or `uri` must be supplied. Supplying both or neither
* returns [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]. See
- * [audio limits](https://cloud.google.com/speech/limits#content).
+ * [content limits](/speech-to-text/quotas#content).
*
*
* Protobuf type {@code google.cloud.speech.v1p1beta1.RecognitionAudio}
@@ -636,7 +638,8 @@ public Builder clearContent() {
/**
*
* URI that points to a file that contains audio data bytes as specified in
- * `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
+ * `RecognitionConfig`. The file must not be compressed (for example, gzip).
+ * Currently, only Google Cloud Storage URIs are
* supported, which must be specified in the following format:
* `gs://bucket_name/object_name` (other URI formats return
* [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
@@ -665,7 +668,8 @@ public java.lang.String getUri() {
/**
*
* URI that points to a file that contains audio data bytes as specified in
- * `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
+ * `RecognitionConfig`. The file must not be compressed (for example, gzip).
+ * Currently, only Google Cloud Storage URIs are
* supported, which must be specified in the following format:
* `gs://bucket_name/object_name` (other URI formats return
* [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
@@ -695,7 +699,8 @@ public java.lang.String getUri() {
/**
*
* URI that points to a file that contains audio data bytes as specified in
- * `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
+ * `RecognitionConfig`. The file must not be compressed (for example, gzip).
+ * Currently, only Google Cloud Storage URIs are
* supported, which must be specified in the following format:
* `gs://bucket_name/object_name` (other URI formats return
* [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
@@ -717,7 +722,8 @@ public Builder setUri(
/**
*
* URI that points to a file that contains audio data bytes as specified in
- * `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
+ * `RecognitionConfig`. The file must not be compressed (for example, gzip).
+ * Currently, only Google Cloud Storage URIs are
* supported, which must be specified in the following format:
* `gs://bucket_name/object_name` (other URI formats return
* [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
@@ -737,7 +743,8 @@ public Builder clearUri() {
/**
*
* URI that points to a file that contains audio data bytes as specified in
- * `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
+ * `RecognitionConfig`. The file must not be compressed (for example, gzip).
+ * Currently, only Google Cloud Storage URIs are
* supported, which must be specified in the following format:
* `gs://bucket_name/object_name` (other URI formats return
* [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionAudioOrBuilder.java b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionAudioOrBuilder.java
index e0127eb4a88a..445c0b83402a 100644
--- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionAudioOrBuilder.java
+++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionAudioOrBuilder.java
@@ -21,7 +21,8 @@ public interface RecognitionAudioOrBuilder extends
/**
*
* URI that points to a file that contains audio data bytes as specified in
- * `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
+ * `RecognitionConfig`. The file must not be compressed (for example, gzip).
+ * Currently, only Google Cloud Storage URIs are
* supported, which must be specified in the following format:
* `gs://bucket_name/object_name` (other URI formats return
* [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
@@ -34,7 +35,8 @@ public interface RecognitionAudioOrBuilder extends
/**
*
* URI that points to a file that contains audio data bytes as specified in
- * `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
+ * `RecognitionConfig`. The file must not be compressed (for example, gzip).
+ * Currently, only Google Cloud Storage URIs are
* supported, which must be specified in the following format:
* `gs://bucket_name/object_name` (other URI formats return
* [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionConfig.java b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionConfig.java
index 16a15615eba4..31952c628671 100644
--- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionConfig.java
+++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionConfig.java
@@ -538,9 +538,9 @@ public int getSampleRateHertz() {
* Valid values for OGG_OPUS are '1'-'254'.
* Valid value for MULAW, AMR, AMR_WB and SPEEX_WITH_HEADER_BYTE is only `1`.
* If `0` or omitted, defaults to one channel (mono).
- * NOTE: We only recognize the first channel by default.
+ * Note: We only recognize the first channel by default.
* To perform independent recognition on each channel set
- * enable_separate_recognition_per_channel to 'true'.
+ * `enable_separate_recognition_per_channel` to 'true'.
*
*
* int32 audio_channel_count = 7;
@@ -553,12 +553,12 @@ public int getAudioChannelCount() {
private boolean enableSeparateRecognitionPerChannel_;
/**
*
- * This needs to be set to ‘true’ explicitly and audio_channel_count > 1
+ * This needs to be set to ‘true’ explicitly and `audio_channel_count` > 1
* to get each channel recognized separately. The recognition result will
- * contain a channel_tag field to state which channel that result belongs to.
- * If this is not ‘true’, we will only recognize the first channel.
- * NOTE: The request is also billed cumulatively for all channels recognized:
- * (audio_channel_count times the audio length)
+ * contain a `channel_tag` field to state which channel that result belongs
+ * to. If this is not true, we will only recognize the first channel. The
+ * request is billed cumulatively for all channels recognized:
+ * `audio_channel_count` multiplied by the length of the audio.
*
*
* bool enable_separate_recognition_per_channel = 12;
@@ -574,7 +574,7 @@ public boolean getEnableSeparateRecognitionPerChannel() {
* *Required* The language of the supplied audio as a
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
* Example: "en-US".
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
*
*
@@ -597,7 +597,7 @@ public java.lang.String getLanguageCode() {
* *Required* The language of the supplied audio as a
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
* Example: "en-US".
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
*
*
@@ -624,13 +624,13 @@ public java.lang.String getLanguageCode() {
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -646,13 +646,13 @@ public java.lang.String getLanguageCode() {
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -667,13 +667,13 @@ public int getAlternativeLanguageCodesCount() {
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -688,13 +688,13 @@ public java.lang.String getAlternativeLanguageCodes(int index) {
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -744,7 +744,9 @@ public boolean getProfanityFilter() {
private java.util.List speechContexts_;
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -754,7 +756,9 @@ public java.util.List getSpeech
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -765,7 +769,9 @@ public java.util.List getSpeech
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -775,7 +781,9 @@ public int getSpeechContextsCount() {
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -785,7 +793,9 @@ public com.google.cloud.speech.v1p1beta1.SpeechContext getSpeechContexts(int ind
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -834,9 +844,9 @@ public boolean getEnableWordConfidence() {
* This feature is only available in select languages. Setting this for
* requests in other languages has no effect at all.
* The default 'false' value does not add punctuation to result hypotheses.
- * NOTE: "This is currently offered as an experimental service, complimentary
+ * Note: This is currently offered as an experimental service, complimentary
* to all users. In the future this may be exclusively available as a
- * premium feature."
+ * premium feature.
*
*
* bool enable_automatic_punctuation = 11;
@@ -853,9 +863,11 @@ public boolean getEnableAutomaticPunctuation() {
* the top alternative of the recognition result using a speaker_tag provided
* in the WordInfo.
* Note: When this is true, we send all the words from the beginning of the
- * audio for the top alternative in every consecutive responses.
+ * audio for the top alternative in every consecutive STREAMING responses.
* This is done in order to improve our speaker tags as our models learn to
* identify the speakers in the conversation over time.
+ * For non-streaming requests, the diarization results will be provided only
+ * in the top alternative of the FINAL SpeechRecognitionResult.
*
*
* bool enable_speaker_diarization = 16;
@@ -1022,13 +1034,16 @@ public java.lang.String getModel() {
/**
*
* *Optional* Set to true to use an enhanced model for speech recognition.
- * You must also set the `model` field to a valid, enhanced model. If
- * `use_enhanced` is set to true and the `model` field is not set, then
- * `use_enhanced` is ignored. If `use_enhanced` is true and an enhanced
- * version of the specified model does not exist, then the speech is
- * recognized using the standard version of the specified model.
- * Enhanced speech models require that you opt-in to the audio logging using
- * instructions in the [alpha documentation](/speech/data-sharing). If you set
+ * If `use_enhanced` is set to true and the `model` field is not set, then
+ * an appropriate enhanced model is chosen if:
+ * 1. project is eligible for requesting enhanced models
+ * 2. an enhanced model exists for the audio
+ * If `use_enhanced` is true and an enhanced version of the specified model
+ * does not exist, then the speech is recognized using the standard version
+ * of the specified model.
+ * Enhanced speech models require that you opt-in to data logging using
+ * instructions in the
+ * [documentation](/speech-to-text/docs/enable-data-logging). If you set
* `use_enhanced` to true and you have not enabled audio logging, then you
* will receive an error.
*
@@ -1837,9 +1852,9 @@ public Builder clearSampleRateHertz() {
* Valid values for OGG_OPUS are '1'-'254'.
* Valid value for MULAW, AMR, AMR_WB and SPEEX_WITH_HEADER_BYTE is only `1`.
* If `0` or omitted, defaults to one channel (mono).
- * NOTE: We only recognize the first channel by default.
+ * Note: We only recognize the first channel by default.
* To perform independent recognition on each channel set
- * enable_separate_recognition_per_channel to 'true'.
+ * `enable_separate_recognition_per_channel` to 'true'.
*
*
* int32 audio_channel_count = 7;
@@ -1855,9 +1870,9 @@ public int getAudioChannelCount() {
* Valid values for OGG_OPUS are '1'-'254'.
* Valid value for MULAW, AMR, AMR_WB and SPEEX_WITH_HEADER_BYTE is only `1`.
* If `0` or omitted, defaults to one channel (mono).
- * NOTE: We only recognize the first channel by default.
+ * Note: We only recognize the first channel by default.
* To perform independent recognition on each channel set
- * enable_separate_recognition_per_channel to 'true'.
+ * `enable_separate_recognition_per_channel` to 'true'.
*
*
* int32 audio_channel_count = 7;
@@ -1876,9 +1891,9 @@ public Builder setAudioChannelCount(int value) {
* Valid values for OGG_OPUS are '1'-'254'.
* Valid value for MULAW, AMR, AMR_WB and SPEEX_WITH_HEADER_BYTE is only `1`.
* If `0` or omitted, defaults to one channel (mono).
- * NOTE: We only recognize the first channel by default.
+ * Note: We only recognize the first channel by default.
* To perform independent recognition on each channel set
- * enable_separate_recognition_per_channel to 'true'.
+ * `enable_separate_recognition_per_channel` to 'true'.
*
*
* int32 audio_channel_count = 7;
@@ -1893,12 +1908,12 @@ public Builder clearAudioChannelCount() {
private boolean enableSeparateRecognitionPerChannel_ ;
/**
*
- * This needs to be set to ‘true’ explicitly and audio_channel_count > 1
+ * This needs to be set to ‘true’ explicitly and `audio_channel_count` > 1
* to get each channel recognized separately. The recognition result will
- * contain a channel_tag field to state which channel that result belongs to.
- * If this is not ‘true’, we will only recognize the first channel.
- * NOTE: The request is also billed cumulatively for all channels recognized:
- * (audio_channel_count times the audio length)
+ * contain a `channel_tag` field to state which channel that result belongs
+ * to. If this is not true, we will only recognize the first channel. The
+ * request is billed cumulatively for all channels recognized:
+ * `audio_channel_count` multiplied by the length of the audio.
*
*
* bool enable_separate_recognition_per_channel = 12;
@@ -1908,12 +1923,12 @@ public boolean getEnableSeparateRecognitionPerChannel() {
}
/**
*
- * This needs to be set to ‘true’ explicitly and audio_channel_count > 1
+ * This needs to be set to ‘true’ explicitly and `audio_channel_count` > 1
* to get each channel recognized separately. The recognition result will
- * contain a channel_tag field to state which channel that result belongs to.
- * If this is not ‘true’, we will only recognize the first channel.
- * NOTE: The request is also billed cumulatively for all channels recognized:
- * (audio_channel_count times the audio length)
+ * contain a `channel_tag` field to state which channel that result belongs
+ * to. If this is not true, we will only recognize the first channel. The
+ * request is billed cumulatively for all channels recognized:
+ * `audio_channel_count` multiplied by the length of the audio.
*
*
* bool enable_separate_recognition_per_channel = 12;
@@ -1926,12 +1941,12 @@ public Builder setEnableSeparateRecognitionPerChannel(boolean value) {
}
/**
*
- * This needs to be set to ‘true’ explicitly and audio_channel_count > 1
+ * This needs to be set to ‘true’ explicitly and `audio_channel_count` > 1
* to get each channel recognized separately. The recognition result will
- * contain a channel_tag field to state which channel that result belongs to.
- * If this is not ‘true’, we will only recognize the first channel.
- * NOTE: The request is also billed cumulatively for all channels recognized:
- * (audio_channel_count times the audio length)
+ * contain a `channel_tag` field to state which channel that result belongs
+ * to. If this is not true, we will only recognize the first channel. The
+ * request is billed cumulatively for all channels recognized:
+ * `audio_channel_count` multiplied by the length of the audio.
*
*
* bool enable_separate_recognition_per_channel = 12;
@@ -1949,7 +1964,7 @@ public Builder clearEnableSeparateRecognitionPerChannel() {
* *Required* The language of the supplied audio as a
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
* Example: "en-US".
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
*
*
@@ -1972,7 +1987,7 @@ public java.lang.String getLanguageCode() {
* *Required* The language of the supplied audio as a
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
* Example: "en-US".
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
*
*
@@ -1996,7 +2011,7 @@ public java.lang.String getLanguageCode() {
* *Required* The language of the supplied audio as a
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
* Example: "en-US".
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
*
*
@@ -2017,7 +2032,7 @@ public Builder setLanguageCode(
* *Required* The language of the supplied audio as a
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
* Example: "en-US".
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
*
*
@@ -2034,7 +2049,7 @@ public Builder clearLanguageCode() {
* *Required* The language of the supplied audio as a
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
* Example: "en-US".
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
*
*
@@ -2064,13 +2079,13 @@ private void ensureAlternativeLanguageCodesIsMutable() {
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -2086,13 +2101,13 @@ private void ensureAlternativeLanguageCodesIsMutable() {
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -2107,13 +2122,13 @@ public int getAlternativeLanguageCodesCount() {
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -2128,13 +2143,13 @@ public java.lang.String getAlternativeLanguageCodes(int index) {
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -2150,13 +2165,13 @@ public java.lang.String getAlternativeLanguageCodes(int index) {
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -2178,13 +2193,13 @@ public Builder setAlternativeLanguageCodes(
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -2206,13 +2221,13 @@ public Builder addAlternativeLanguageCodes(
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -2232,13 +2247,13 @@ public Builder addAllAlternativeLanguageCodes(
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -2256,13 +2271,13 @@ public Builder clearAlternativeLanguageCodes() {
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -2395,7 +2410,9 @@ private void ensureSpeechContextsIsMutable() {
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2409,7 +2426,9 @@ public java.util.List
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2423,7 +2442,9 @@ public int getSpeechContextsCount() {
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2437,7 +2458,9 @@ public com.google.cloud.speech.v1p1beta1.SpeechContext getSpeechContexts(int ind
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2458,7 +2481,9 @@ public Builder setSpeechContexts(
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2476,7 +2501,9 @@ public Builder setSpeechContexts(
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2496,7 +2523,9 @@ public Builder addSpeechContexts(com.google.cloud.speech.v1p1beta1.SpeechContext
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2517,7 +2546,9 @@ public Builder addSpeechContexts(
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2535,7 +2566,9 @@ public Builder addSpeechContexts(
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2553,7 +2586,9 @@ public Builder addSpeechContexts(
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2572,7 +2607,9 @@ public Builder addAllSpeechContexts(
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2589,7 +2626,9 @@ public Builder clearSpeechContexts() {
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2606,7 +2645,9 @@ public Builder removeSpeechContexts(int index) {
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2617,7 +2658,9 @@ public com.google.cloud.speech.v1p1beta1.SpeechContext.Builder getSpeechContexts
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2631,7 +2674,9 @@ public com.google.cloud.speech.v1p1beta1.SpeechContextOrBuilder getSpeechContext
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2646,7 +2691,9 @@ public com.google.cloud.speech.v1p1beta1.SpeechContextOrBuilder getSpeechContext
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2657,7 +2704,9 @@ public com.google.cloud.speech.v1p1beta1.SpeechContext.Builder addSpeechContexts
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2669,7 +2718,9 @@ public com.google.cloud.speech.v1p1beta1.SpeechContext.Builder addSpeechContexts
}
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -2791,9 +2842,9 @@ public Builder clearEnableWordConfidence() {
* This feature is only available in select languages. Setting this for
* requests in other languages has no effect at all.
* The default 'false' value does not add punctuation to result hypotheses.
- * NOTE: "This is currently offered as an experimental service, complimentary
+ * Note: This is currently offered as an experimental service, complimentary
* to all users. In the future this may be exclusively available as a
- * premium feature."
+ * premium feature.
*
*
* bool enable_automatic_punctuation = 11;
@@ -2807,9 +2858,9 @@ public boolean getEnableAutomaticPunctuation() {
* This feature is only available in select languages. Setting this for
* requests in other languages has no effect at all.
* The default 'false' value does not add punctuation to result hypotheses.
- * NOTE: "This is currently offered as an experimental service, complimentary
+ * Note: This is currently offered as an experimental service, complimentary
* to all users. In the future this may be exclusively available as a
- * premium feature."
+ * premium feature.
*
*
* bool enable_automatic_punctuation = 11;
@@ -2826,9 +2877,9 @@ public Builder setEnableAutomaticPunctuation(boolean value) {
* This feature is only available in select languages. Setting this for
* requests in other languages has no effect at all.
* The default 'false' value does not add punctuation to result hypotheses.
- * NOTE: "This is currently offered as an experimental service, complimentary
+ * Note: This is currently offered as an experimental service, complimentary
* to all users. In the future this may be exclusively available as a
- * premium feature."
+ * premium feature.
*
*
* bool enable_automatic_punctuation = 11;
@@ -2847,9 +2898,11 @@ public Builder clearEnableAutomaticPunctuation() {
* the top alternative of the recognition result using a speaker_tag provided
* in the WordInfo.
* Note: When this is true, we send all the words from the beginning of the
- * audio for the top alternative in every consecutive responses.
+ * audio for the top alternative in every consecutive STREAMING responses.
* This is done in order to improve our speaker tags as our models learn to
* identify the speakers in the conversation over time.
+ * For non-streaming requests, the diarization results will be provided only
+ * in the top alternative of the FINAL SpeechRecognitionResult.
*
*
* bool enable_speaker_diarization = 16;
@@ -2863,9 +2916,11 @@ public boolean getEnableSpeakerDiarization() {
* the top alternative of the recognition result using a speaker_tag provided
* in the WordInfo.
* Note: When this is true, we send all the words from the beginning of the
- * audio for the top alternative in every consecutive responses.
+ * audio for the top alternative in every consecutive STREAMING responses.
* This is done in order to improve our speaker tags as our models learn to
* identify the speakers in the conversation over time.
+ * For non-streaming requests, the diarization results will be provided only
+ * in the top alternative of the FINAL SpeechRecognitionResult.
*
*
* bool enable_speaker_diarization = 16;
@@ -2882,9 +2937,11 @@ public Builder setEnableSpeakerDiarization(boolean value) {
* the top alternative of the recognition result using a speaker_tag provided
* in the WordInfo.
* Note: When this is true, we send all the words from the beginning of the
- * audio for the top alternative in every consecutive responses.
+ * audio for the top alternative in every consecutive STREAMING responses.
* This is done in order to improve our speaker tags as our models learn to
* identify the speakers in the conversation over time.
+ * For non-streaming requests, the diarization results will be provided only
+ * in the top alternative of the FINAL SpeechRecognitionResult.
*
*
* bool enable_speaker_diarization = 16;
@@ -3344,13 +3401,16 @@ public Builder setModelBytes(
/**
*
* *Optional* Set to true to use an enhanced model for speech recognition.
- * You must also set the `model` field to a valid, enhanced model. If
- * `use_enhanced` is set to true and the `model` field is not set, then
- * `use_enhanced` is ignored. If `use_enhanced` is true and an enhanced
- * version of the specified model does not exist, then the speech is
- * recognized using the standard version of the specified model.
- * Enhanced speech models require that you opt-in to the audio logging using
- * instructions in the [alpha documentation](/speech/data-sharing). If you set
+ * If `use_enhanced` is set to true and the `model` field is not set, then
+ * an appropriate enhanced model is chosen if:
+ * 1. project is eligible for requesting enhanced models
+ * 2. an enhanced model exists for the audio
+ * If `use_enhanced` is true and an enhanced version of the specified model
+ * does not exist, then the speech is recognized using the standard version
+ * of the specified model.
+ * Enhanced speech models require that you opt-in to data logging using
+ * instructions in the
+ * [documentation](/speech-to-text/docs/enable-data-logging). If you set
* `use_enhanced` to true and you have not enabled audio logging, then you
* will receive an error.
*
@@ -3363,13 +3423,16 @@ public boolean getUseEnhanced() {
/**
*
* *Optional* Set to true to use an enhanced model for speech recognition.
- * You must also set the `model` field to a valid, enhanced model. If
- * `use_enhanced` is set to true and the `model` field is not set, then
- * `use_enhanced` is ignored. If `use_enhanced` is true and an enhanced
- * version of the specified model does not exist, then the speech is
- * recognized using the standard version of the specified model.
- * Enhanced speech models require that you opt-in to the audio logging using
- * instructions in the [alpha documentation](/speech/data-sharing). If you set
+ * If `use_enhanced` is set to true and the `model` field is not set, then
+ * an appropriate enhanced model is chosen if:
+ * 1. project is eligible for requesting enhanced models
+ * 2. an enhanced model exists for the audio
+ * If `use_enhanced` is true and an enhanced version of the specified model
+ * does not exist, then the speech is recognized using the standard version
+ * of the specified model.
+ * Enhanced speech models require that you opt-in to data logging using
+ * instructions in the
+ * [documentation](/speech-to-text/docs/enable-data-logging). If you set
* `use_enhanced` to true and you have not enabled audio logging, then you
* will receive an error.
*
@@ -3385,13 +3448,16 @@ public Builder setUseEnhanced(boolean value) {
/**
*
* *Optional* Set to true to use an enhanced model for speech recognition.
- * You must also set the `model` field to a valid, enhanced model. If
- * `use_enhanced` is set to true and the `model` field is not set, then
- * `use_enhanced` is ignored. If `use_enhanced` is true and an enhanced
- * version of the specified model does not exist, then the speech is
- * recognized using the standard version of the specified model.
- * Enhanced speech models require that you opt-in to the audio logging using
- * instructions in the [alpha documentation](/speech/data-sharing). If you set
+ * If `use_enhanced` is set to true and the `model` field is not set, then
+ * an appropriate enhanced model is chosen if:
+ * 1. project is eligible for requesting enhanced models
+ * 2. an enhanced model exists for the audio
+ * If `use_enhanced` is true and an enhanced version of the specified model
+ * does not exist, then the speech is recognized using the standard version
+ * of the specified model.
+ * Enhanced speech models require that you opt-in to data logging using
+ * instructions in the
+ * [documentation](/speech-to-text/docs/enable-data-logging). If you set
* `use_enhanced` to true and you have not enabled audio logging, then you
* will receive an error.
*
diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionConfigOrBuilder.java b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionConfigOrBuilder.java
index 31216fcceb26..bc1c1e942597 100644
--- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionConfigOrBuilder.java
+++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/RecognitionConfigOrBuilder.java
@@ -51,9 +51,9 @@ public interface RecognitionConfigOrBuilder extends
* Valid values for OGG_OPUS are '1'-'254'.
* Valid value for MULAW, AMR, AMR_WB and SPEEX_WITH_HEADER_BYTE is only `1`.
* If `0` or omitted, defaults to one channel (mono).
- * NOTE: We only recognize the first channel by default.
+ * Note: We only recognize the first channel by default.
* To perform independent recognition on each channel set
- * enable_separate_recognition_per_channel to 'true'.
+ * `enable_separate_recognition_per_channel` to 'true'.
*
*
* int32 audio_channel_count = 7;
@@ -62,12 +62,12 @@ public interface RecognitionConfigOrBuilder extends
/**
*
- * This needs to be set to ‘true’ explicitly and audio_channel_count > 1
+ * This needs to be set to ‘true’ explicitly and `audio_channel_count` > 1
* to get each channel recognized separately. The recognition result will
- * contain a channel_tag field to state which channel that result belongs to.
- * If this is not ‘true’, we will only recognize the first channel.
- * NOTE: The request is also billed cumulatively for all channels recognized:
- * (audio_channel_count times the audio length)
+ * contain a `channel_tag` field to state which channel that result belongs
+ * to. If this is not true, we will only recognize the first channel. The
+ * request is billed cumulatively for all channels recognized:
+ * `audio_channel_count` multiplied by the length of the audio.
*
*
* bool enable_separate_recognition_per_channel = 12;
@@ -79,7 +79,7 @@ public interface RecognitionConfigOrBuilder extends
* *Required* The language of the supplied audio as a
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
* Example: "en-US".
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
*
*
@@ -91,7 +91,7 @@ public interface RecognitionConfigOrBuilder extends
* *Required* The language of the supplied audio as a
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
* Example: "en-US".
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
*
*
@@ -105,13 +105,13 @@ public interface RecognitionConfigOrBuilder extends
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -125,13 +125,13 @@ public interface RecognitionConfigOrBuilder extends
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -144,13 +144,13 @@ public interface RecognitionConfigOrBuilder extends
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -163,13 +163,13 @@ public interface RecognitionConfigOrBuilder extends
* *Optional* A list of up to 3 additional
* [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
* listing possible alternative languages of the supplied audio.
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
+ * See [Language Support](/speech-to-text/docs/languages)
* for a list of the currently supported language codes.
* If alternative languages are listed, recognition result will contain
* recognition in the most likely language detected including the main
* language_code. The recognition result will include the language tag
* of the language detected in the audio.
- * NOTE: This feature is only supported for Voice Command and Voice Search
+ * Note: This feature is only supported for Voice Command and Voice Search
* use cases and performance may vary for other use cases (e.g., phone call
* transcription).
*
@@ -207,7 +207,9 @@ public interface RecognitionConfigOrBuilder extends
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -216,7 +218,9 @@ public interface RecognitionConfigOrBuilder extends
getSpeechContextsList();
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -224,7 +228,9 @@ public interface RecognitionConfigOrBuilder extends
com.google.cloud.speech.v1p1beta1.SpeechContext getSpeechContexts(int index);
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -232,7 +238,9 @@ public interface RecognitionConfigOrBuilder extends
int getSpeechContextsCount();
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -241,7 +249,9 @@ public interface RecognitionConfigOrBuilder extends
getSpeechContextsOrBuilderList();
/**
*
- * *Optional* A means to provide context to assist the speech recognition.
+ * *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ * A means to provide context to assist the speech recognition. For more
+ * information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
*
*
* repeated .google.cloud.speech.v1p1beta1.SpeechContext speech_contexts = 6;
@@ -278,9 +288,9 @@ com.google.cloud.speech.v1p1beta1.SpeechContextOrBuilder getSpeechContextsOrBuil
* This feature is only available in select languages. Setting this for
* requests in other languages has no effect at all.
* The default 'false' value does not add punctuation to result hypotheses.
- * NOTE: "This is currently offered as an experimental service, complimentary
+ * Note: This is currently offered as an experimental service, complimentary
* to all users. In the future this may be exclusively available as a
- * premium feature."
+ * premium feature.
*
*
* bool enable_automatic_punctuation = 11;
@@ -293,9 +303,11 @@ com.google.cloud.speech.v1p1beta1.SpeechContextOrBuilder getSpeechContextsOrBuil
* the top alternative of the recognition result using a speaker_tag provided
* in the WordInfo.
* Note: When this is true, we send all the words from the beginning of the
- * audio for the top alternative in every consecutive responses.
+ * audio for the top alternative in every consecutive STREAMING responses.
* This is done in order to improve our speaker tags as our models learn to
* identify the speakers in the conversation over time.
+ * For non-streaming requests, the diarization results will be provided only
+ * in the top alternative of the FINAL SpeechRecognitionResult.
*
*
* bool enable_speaker_diarization = 16;
@@ -422,13 +434,16 @@ com.google.cloud.speech.v1p1beta1.SpeechContextOrBuilder getSpeechContextsOrBuil
/**
*
* *Optional* Set to true to use an enhanced model for speech recognition.
- * You must also set the `model` field to a valid, enhanced model. If
- * `use_enhanced` is set to true and the `model` field is not set, then
- * `use_enhanced` is ignored. If `use_enhanced` is true and an enhanced
- * version of the specified model does not exist, then the speech is
- * recognized using the standard version of the specified model.
- * Enhanced speech models require that you opt-in to the audio logging using
- * instructions in the [alpha documentation](/speech/data-sharing). If you set
+ * If `use_enhanced` is set to true and the `model` field is not set, then
+ * an appropriate enhanced model is chosen if:
+ * 1. project is eligible for requesting enhanced models
+ * 2. an enhanced model exists for the audio
+ * If `use_enhanced` is true and an enhanced version of the specified model
+ * does not exist, then the speech is recognized using the standard version
+ * of the specified model.
+ * Enhanced speech models require that you opt-in to data logging using
+ * instructions in the
+ * [documentation](/speech-to-text/docs/enable-data-logging). If you set
* `use_enhanced` to true and you have not enabled audio logging, then you
* will receive an error.
*
diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechContext.java b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechContext.java
index 88bf669987cb..1b0569d7f588 100644
--- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechContext.java
+++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechContext.java
@@ -101,7 +101,7 @@ private SpeechContext(
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -117,7 +117,7 @@ private SpeechContext(
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -132,7 +132,7 @@ public int getPhrasesCount() {
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -147,7 +147,7 @@ public java.lang.String getPhrases(int index) {
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -498,7 +498,7 @@ private void ensurePhrasesIsMutable() {
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -514,7 +514,7 @@ private void ensurePhrasesIsMutable() {
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -529,7 +529,7 @@ public int getPhrasesCount() {
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -544,7 +544,7 @@ public java.lang.String getPhrases(int index) {
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -560,7 +560,7 @@ public java.lang.String getPhrases(int index) {
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -582,7 +582,7 @@ public Builder setPhrases(
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -604,7 +604,7 @@ public Builder addPhrases(
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -624,7 +624,7 @@ public Builder addAllPhrases(
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -642,7 +642,7 @@ public Builder clearPhrases() {
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechContextOrBuilder.java b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechContextOrBuilder.java
index 3d5250046cf9..f9ddb6724147 100644
--- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechContextOrBuilder.java
+++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechContextOrBuilder.java
@@ -14,7 +14,7 @@ public interface SpeechContextOrBuilder extends
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -28,7 +28,7 @@ public interface SpeechContextOrBuilder extends
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -41,7 +41,7 @@ public interface SpeechContextOrBuilder extends
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
@@ -54,7 +54,7 @@ public interface SpeechContextOrBuilder extends
* to improve the accuracy for specific words and phrases, for example, if
* specific commands are typically spoken by the user. This can also be used
* to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
+ * [usage limits](/speech-to-text/quotas#content).
*
*
* repeated string phrases = 1;
diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechProto.java b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechProto.java
index f64d1ab6b196..85229ebb9003 100644
--- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechProto.java
+++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechProto.java
@@ -108,123 +108,125 @@ public static void registerAllExtensions(
"1\032\034google/api/annotations.proto\032#google/" +
"longrunning/operations.proto\032\031google/pro" +
"tobuf/any.proto\032\036google/protobuf/duratio" +
- "n.proto\032\037google/protobuf/timestamp.proto" +
- "\032\027google/rpc/status.proto\"\224\001\n\020RecognizeR" +
- "equest\022@\n\006config\030\001 \001(\01320.google.cloud.sp" +
- "eech.v1p1beta1.RecognitionConfig\022>\n\005audi" +
- "o\030\002 \001(\0132/.google.cloud.speech.v1p1beta1." +
- "RecognitionAudio\"\237\001\n\033LongRunningRecogniz" +
- "eRequest\022@\n\006config\030\001 \001(\01320.google.cloud." +
- "speech.v1p1beta1.RecognitionConfig\022>\n\005au" +
- "dio\030\002 \001(\0132/.google.cloud.speech.v1p1beta" +
- "1.RecognitionAudio\"\240\001\n\031StreamingRecogniz" +
- "eRequest\022U\n\020streaming_config\030\001 \001(\01329.goo" +
- "gle.cloud.speech.v1p1beta1.StreamingReco" +
- "gnitionConfigH\000\022\027\n\raudio_content\030\002 \001(\014H\000" +
- "B\023\n\021streaming_request\"\221\001\n\032StreamingRecog" +
- "nitionConfig\022@\n\006config\030\001 \001(\01320.google.cl" +
- "oud.speech.v1p1beta1.RecognitionConfig\022\030" +
- "\n\020single_utterance\030\002 \001(\010\022\027\n\017interim_resu" +
- "lts\030\003 \001(\010\"\254\006\n\021RecognitionConfig\022P\n\010encod" +
- "ing\030\001 \001(\0162>.google.cloud.speech.v1p1beta" +
- "1.RecognitionConfig.AudioEncoding\022\031\n\021sam" +
- "ple_rate_hertz\030\002 \001(\005\022\033\n\023audio_channel_co" +
- "unt\030\007 \001(\005\022/\n\'enable_separate_recognition" +
- "_per_channel\030\014 \001(\010\022\025\n\rlanguage_code\030\003 \001(" +
- "\t\022\"\n\032alternative_language_codes\030\022 \003(\t\022\030\n" +
- "\020max_alternatives\030\004 \001(\005\022\030\n\020profanity_fil" +
- "ter\030\005 \001(\010\022E\n\017speech_contexts\030\006 \003(\0132,.goo" +
- "gle.cloud.speech.v1p1beta1.SpeechContext" +
- "\022 \n\030enable_word_time_offsets\030\010 \001(\010\022\036\n\026en" +
- "able_word_confidence\030\017 \001(\010\022$\n\034enable_aut" +
- "omatic_punctuation\030\013 \001(\010\022\"\n\032enable_speak" +
- "er_diarization\030\020 \001(\010\022!\n\031diarization_spea" +
- "ker_count\030\021 \001(\005\022D\n\010metadata\030\t \001(\01322.goog" +
- "le.cloud.speech.v1p1beta1.RecognitionMet" +
- "adata\022\r\n\005model\030\r \001(\t\022\024\n\014use_enhanced\030\016 \001" +
- "(\010\"\213\001\n\rAudioEncoding\022\030\n\024ENCODING_UNSPECI" +
- "FIED\020\000\022\014\n\010LINEAR16\020\001\022\010\n\004FLAC\020\002\022\t\n\005MULAW\020" +
- "\003\022\007\n\003AMR\020\004\022\n\n\006AMR_WB\020\005\022\014\n\010OGG_OPUS\020\006\022\032\n\026" +
- "SPEEX_WITH_HEADER_BYTE\020\007\"\323\010\n\023Recognition" +
- "Metadata\022\\\n\020interaction_type\030\001 \001(\0162B.goo" +
- "gle.cloud.speech.v1p1beta1.RecognitionMe" +
- "tadata.InteractionType\022$\n\034industry_naics" +
- "_code_of_audio\030\003 \001(\r\022b\n\023microphone_dista" +
- "nce\030\004 \001(\0162E.google.cloud.speech.v1p1beta" +
- "1.RecognitionMetadata.MicrophoneDistance" +
- "\022a\n\023original_media_type\030\005 \001(\0162D.google.c" +
- "loud.speech.v1p1beta1.RecognitionMetadat" +
- "a.OriginalMediaType\022e\n\025recording_device_" +
- "type\030\006 \001(\0162F.google.cloud.speech.v1p1bet" +
- "a1.RecognitionMetadata.RecordingDeviceTy" +
- "pe\022\035\n\025recording_device_name\030\007 \001(\t\022\032\n\022ori" +
- "ginal_mime_type\030\010 \001(\t\022\025\n\robfuscated_id\030\t" +
- " \001(\003\022\023\n\013audio_topic\030\n \001(\t\"\305\001\n\017Interactio" +
- "nType\022 \n\034INTERACTION_TYPE_UNSPECIFIED\020\000\022" +
- "\016\n\nDISCUSSION\020\001\022\020\n\014PRESENTATION\020\002\022\016\n\nPHO" +
- "NE_CALL\020\003\022\r\n\tVOICEMAIL\020\004\022\033\n\027PROFESSIONAL" +
- "LY_PRODUCED\020\005\022\020\n\014VOICE_SEARCH\020\006\022\021\n\rVOICE" +
- "_COMMAND\020\007\022\r\n\tDICTATION\020\010\"d\n\022MicrophoneD" +
- "istance\022#\n\037MICROPHONE_DISTANCE_UNSPECIFI" +
- "ED\020\000\022\r\n\tNEARFIELD\020\001\022\014\n\010MIDFIELD\020\002\022\014\n\010FAR" +
- "FIELD\020\003\"N\n\021OriginalMediaType\022#\n\037ORIGINAL" +
- "_MEDIA_TYPE_UNSPECIFIED\020\000\022\t\n\005AUDIO\020\001\022\t\n\005" +
- "VIDEO\020\002\"\244\001\n\023RecordingDeviceType\022%\n!RECOR" +
- "DING_DEVICE_TYPE_UNSPECIFIED\020\000\022\016\n\nSMARTP" +
- "HONE\020\001\022\006\n\002PC\020\002\022\016\n\nPHONE_LINE\020\003\022\013\n\007VEHICL" +
- "E\020\004\022\030\n\024OTHER_OUTDOOR_DEVICE\020\005\022\027\n\023OTHER_I" +
- "NDOOR_DEVICE\020\006\" \n\rSpeechContext\022\017\n\007phras" +
- "es\030\001 \003(\t\"D\n\020RecognitionAudio\022\021\n\007content\030" +
- "\001 \001(\014H\000\022\r\n\003uri\030\002 \001(\tH\000B\016\n\014audio_source\"\\" +
- "\n\021RecognizeResponse\022G\n\007results\030\002 \003(\01326.g" +
- "oogle.cloud.speech.v1p1beta1.SpeechRecog" +
- "nitionResult\"g\n\034LongRunningRecognizeResp" +
- "onse\022G\n\007results\030\002 \003(\01326.google.cloud.spe" +
- "ech.v1p1beta1.SpeechRecognitionResult\"\236\001" +
- "\n\034LongRunningRecognizeMetadata\022\030\n\020progre" +
- "ss_percent\030\001 \001(\005\022.\n\nstart_time\030\002 \001(\0132\032.g" +
- "oogle.protobuf.Timestamp\0224\n\020last_update_" +
- "time\030\003 \001(\0132\032.google.protobuf.Timestamp\"\277" +
- "\002\n\032StreamingRecognizeResponse\022!\n\005error\030\001" +
- " \001(\0132\022.google.rpc.Status\022J\n\007results\030\002 \003(" +
- "\01329.google.cloud.speech.v1p1beta1.Stream" +
- "ingRecognitionResult\022d\n\021speech_event_typ" +
- "e\030\004 \001(\0162I.google.cloud.speech.v1p1beta1." +
- "StreamingRecognizeResponse.SpeechEventTy" +
- "pe\"L\n\017SpeechEventType\022\034\n\030SPEECH_EVENT_UN" +
- "SPECIFIED\020\000\022\033\n\027END_OF_SINGLE_UTTERANCE\020\001" +
- "\"\300\001\n\032StreamingRecognitionResult\022Q\n\014alter" +
- "natives\030\001 \003(\0132;.google.cloud.speech.v1p1" +
- "beta1.SpeechRecognitionAlternative\022\020\n\010is" +
- "_final\030\002 \001(\010\022\021\n\tstability\030\003 \001(\002\022\023\n\013chann" +
- "el_tag\030\005 \001(\005\022\025\n\rlanguage_code\030\006 \001(\t\"\230\001\n\027" +
- "SpeechRecognitionResult\022Q\n\014alternatives\030" +
- "\001 \003(\0132;.google.cloud.speech.v1p1beta1.Sp" +
- "eechRecognitionAlternative\022\023\n\013channel_ta" +
- "g\030\002 \001(\005\022\025\n\rlanguage_code\030\005 \001(\t\"~\n\034Speech" +
- "RecognitionAlternative\022\022\n\ntranscript\030\001 \001" +
- "(\t\022\022\n\nconfidence\030\002 \001(\002\0226\n\005words\030\003 \003(\0132\'." +
- "google.cloud.speech.v1p1beta1.WordInfo\"\235" +
- "\001\n\010WordInfo\022-\n\nstart_time\030\001 \001(\0132\031.google" +
- ".protobuf.Duration\022+\n\010end_time\030\002 \001(\0132\031.g" +
- "oogle.protobuf.Duration\022\014\n\004word\030\003 \001(\t\022\022\n" +
- "\nconfidence\030\004 \001(\002\022\023\n\013speaker_tag\030\005 \001(\0052\330" +
- "\003\n\006Speech\022\226\001\n\tRecognize\022/.google.cloud.s" +
- "peech.v1p1beta1.RecognizeRequest\0320.googl" +
- "e.cloud.speech.v1p1beta1.RecognizeRespon" +
- "se\"&\202\323\344\223\002 \"\033/v1p1beta1/speech:recognize:" +
- "\001*\022\244\001\n\024LongRunningRecognize\022:.google.clo" +
- "ud.speech.v1p1beta1.LongRunningRecognize" +
- "Request\032\035.google.longrunning.Operation\"1" +
- "\202\323\344\223\002+\"&/v1p1beta1/speech:longrunningrec" +
- "ognize:\001*\022\215\001\n\022StreamingRecognize\0228.googl" +
- "e.cloud.speech.v1p1beta1.StreamingRecogn" +
- "izeRequest\0329.google.cloud.speech.v1p1bet" +
- "a1.StreamingRecognizeResponse(\0010\001Bz\n!com" +
- ".google.cloud.speech.v1p1beta1B\013SpeechPr" +
- "otoP\001ZCgoogle.golang.org/genproto/google" +
- "apis/cloud/speech/v1p1beta1;speech\370\001\001b\006p" +
- "roto3"
+ "n.proto\032\033google/protobuf/empty.proto\032\037go" +
+ "ogle/protobuf/timestamp.proto\032\027google/rp" +
+ "c/status.proto\"\224\001\n\020RecognizeRequest\022@\n\006c" +
+ "onfig\030\001 \001(\01320.google.cloud.speech.v1p1be" +
+ "ta1.RecognitionConfig\022>\n\005audio\030\002 \001(\0132/.g" +
+ "oogle.cloud.speech.v1p1beta1.Recognition" +
+ "Audio\"\237\001\n\033LongRunningRecognizeRequest\022@\n" +
+ "\006config\030\001 \001(\01320.google.cloud.speech.v1p1" +
+ "beta1.RecognitionConfig\022>\n\005audio\030\002 \001(\0132/" +
+ ".google.cloud.speech.v1p1beta1.Recogniti" +
+ "onAudio\"\240\001\n\031StreamingRecognizeRequest\022U\n" +
+ "\020streaming_config\030\001 \001(\01329.google.cloud.s" +
+ "peech.v1p1beta1.StreamingRecognitionConf" +
+ "igH\000\022\027\n\raudio_content\030\002 \001(\014H\000B\023\n\021streami" +
+ "ng_request\"\221\001\n\032StreamingRecognitionConfi" +
+ "g\022@\n\006config\030\001 \001(\01320.google.cloud.speech." +
+ "v1p1beta1.RecognitionConfig\022\030\n\020single_ut" +
+ "terance\030\002 \001(\010\022\027\n\017interim_results\030\003 \001(\010\"\254" +
+ "\006\n\021RecognitionConfig\022P\n\010encoding\030\001 \001(\0162>" +
+ ".google.cloud.speech.v1p1beta1.Recogniti" +
+ "onConfig.AudioEncoding\022\031\n\021sample_rate_he" +
+ "rtz\030\002 \001(\005\022\033\n\023audio_channel_count\030\007 \001(\005\022/" +
+ "\n\'enable_separate_recognition_per_channe" +
+ "l\030\014 \001(\010\022\025\n\rlanguage_code\030\003 \001(\t\022\"\n\032altern" +
+ "ative_language_codes\030\022 \003(\t\022\030\n\020max_altern" +
+ "atives\030\004 \001(\005\022\030\n\020profanity_filter\030\005 \001(\010\022E" +
+ "\n\017speech_contexts\030\006 \003(\0132,.google.cloud.s" +
+ "peech.v1p1beta1.SpeechContext\022 \n\030enable_" +
+ "word_time_offsets\030\010 \001(\010\022\036\n\026enable_word_c" +
+ "onfidence\030\017 \001(\010\022$\n\034enable_automatic_punc" +
+ "tuation\030\013 \001(\010\022\"\n\032enable_speaker_diarizat" +
+ "ion\030\020 \001(\010\022!\n\031diarization_speaker_count\030\021" +
+ " \001(\005\022D\n\010metadata\030\t \001(\01322.google.cloud.sp" +
+ "eech.v1p1beta1.RecognitionMetadata\022\r\n\005mo" +
+ "del\030\r \001(\t\022\024\n\014use_enhanced\030\016 \001(\010\"\213\001\n\rAudi" +
+ "oEncoding\022\030\n\024ENCODING_UNSPECIFIED\020\000\022\014\n\010L" +
+ "INEAR16\020\001\022\010\n\004FLAC\020\002\022\t\n\005MULAW\020\003\022\007\n\003AMR\020\004\022" +
+ "\n\n\006AMR_WB\020\005\022\014\n\010OGG_OPUS\020\006\022\032\n\026SPEEX_WITH_" +
+ "HEADER_BYTE\020\007\"\323\010\n\023RecognitionMetadata\022\\\n" +
+ "\020interaction_type\030\001 \001(\0162B.google.cloud.s" +
+ "peech.v1p1beta1.RecognitionMetadata.Inte" +
+ "ractionType\022$\n\034industry_naics_code_of_au" +
+ "dio\030\003 \001(\r\022b\n\023microphone_distance\030\004 \001(\0162E" +
+ ".google.cloud.speech.v1p1beta1.Recogniti" +
+ "onMetadata.MicrophoneDistance\022a\n\023origina" +
+ "l_media_type\030\005 \001(\0162D.google.cloud.speech" +
+ ".v1p1beta1.RecognitionMetadata.OriginalM" +
+ "ediaType\022e\n\025recording_device_type\030\006 \001(\0162" +
+ "F.google.cloud.speech.v1p1beta1.Recognit" +
+ "ionMetadata.RecordingDeviceType\022\035\n\025recor" +
+ "ding_device_name\030\007 \001(\t\022\032\n\022original_mime_" +
+ "type\030\010 \001(\t\022\025\n\robfuscated_id\030\t \001(\003\022\023\n\013aud" +
+ "io_topic\030\n \001(\t\"\305\001\n\017InteractionType\022 \n\034IN" +
+ "TERACTION_TYPE_UNSPECIFIED\020\000\022\016\n\nDISCUSSI" +
+ "ON\020\001\022\020\n\014PRESENTATION\020\002\022\016\n\nPHONE_CALL\020\003\022\r" +
+ "\n\tVOICEMAIL\020\004\022\033\n\027PROFESSIONALLY_PRODUCED" +
+ "\020\005\022\020\n\014VOICE_SEARCH\020\006\022\021\n\rVOICE_COMMAND\020\007\022" +
+ "\r\n\tDICTATION\020\010\"d\n\022MicrophoneDistance\022#\n\037" +
+ "MICROPHONE_DISTANCE_UNSPECIFIED\020\000\022\r\n\tNEA" +
+ "RFIELD\020\001\022\014\n\010MIDFIELD\020\002\022\014\n\010FARFIELD\020\003\"N\n\021" +
+ "OriginalMediaType\022#\n\037ORIGINAL_MEDIA_TYPE" +
+ "_UNSPECIFIED\020\000\022\t\n\005AUDIO\020\001\022\t\n\005VIDEO\020\002\"\244\001\n" +
+ "\023RecordingDeviceType\022%\n!RECORDING_DEVICE" +
+ "_TYPE_UNSPECIFIED\020\000\022\016\n\nSMARTPHONE\020\001\022\006\n\002P" +
+ "C\020\002\022\016\n\nPHONE_LINE\020\003\022\013\n\007VEHICLE\020\004\022\030\n\024OTHE" +
+ "R_OUTDOOR_DEVICE\020\005\022\027\n\023OTHER_INDOOR_DEVIC" +
+ "E\020\006\" \n\rSpeechContext\022\017\n\007phrases\030\001 \003(\t\"D\n" +
+ "\020RecognitionAudio\022\021\n\007content\030\001 \001(\014H\000\022\r\n\003" +
+ "uri\030\002 \001(\tH\000B\016\n\014audio_source\"\\\n\021Recognize" +
+ "Response\022G\n\007results\030\002 \003(\01326.google.cloud" +
+ ".speech.v1p1beta1.SpeechRecognitionResul" +
+ "t\"g\n\034LongRunningRecognizeResponse\022G\n\007res" +
+ "ults\030\002 \003(\01326.google.cloud.speech.v1p1bet" +
+ "a1.SpeechRecognitionResult\"\236\001\n\034LongRunni" +
+ "ngRecognizeMetadata\022\030\n\020progress_percent\030" +
+ "\001 \001(\005\022.\n\nstart_time\030\002 \001(\0132\032.google.proto" +
+ "buf.Timestamp\0224\n\020last_update_time\030\003 \001(\0132" +
+ "\032.google.protobuf.Timestamp\"\277\002\n\032Streamin" +
+ "gRecognizeResponse\022!\n\005error\030\001 \001(\0132\022.goog" +
+ "le.rpc.Status\022J\n\007results\030\002 \003(\01329.google." +
+ "cloud.speech.v1p1beta1.StreamingRecognit" +
+ "ionResult\022d\n\021speech_event_type\030\004 \001(\0162I.g" +
+ "oogle.cloud.speech.v1p1beta1.StreamingRe" +
+ "cognizeResponse.SpeechEventType\"L\n\017Speec" +
+ "hEventType\022\034\n\030SPEECH_EVENT_UNSPECIFIED\020\000" +
+ "\022\033\n\027END_OF_SINGLE_UTTERANCE\020\001\"\364\001\n\032Stream" +
+ "ingRecognitionResult\022Q\n\014alternatives\030\001 \003" +
+ "(\0132;.google.cloud.speech.v1p1beta1.Speec" +
+ "hRecognitionAlternative\022\020\n\010is_final\030\002 \001(" +
+ "\010\022\021\n\tstability\030\003 \001(\002\0222\n\017result_end_time\030" +
+ "\004 \001(\0132\031.google.protobuf.Duration\022\023\n\013chan" +
+ "nel_tag\030\005 \001(\005\022\025\n\rlanguage_code\030\006 \001(\t\"\230\001\n" +
+ "\027SpeechRecognitionResult\022Q\n\014alternatives" +
+ "\030\001 \003(\0132;.google.cloud.speech.v1p1beta1.S" +
+ "peechRecognitionAlternative\022\023\n\013channel_t" +
+ "ag\030\002 \001(\005\022\025\n\rlanguage_code\030\005 \001(\t\"~\n\034Speec" +
+ "hRecognitionAlternative\022\022\n\ntranscript\030\001 " +
+ "\001(\t\022\022\n\nconfidence\030\002 \001(\002\0226\n\005words\030\003 \003(\0132\'" +
+ ".google.cloud.speech.v1p1beta1.WordInfo\"" +
+ "\235\001\n\010WordInfo\022-\n\nstart_time\030\001 \001(\0132\031.googl" +
+ "e.protobuf.Duration\022+\n\010end_time\030\002 \001(\0132\031." +
+ "google.protobuf.Duration\022\014\n\004word\030\003 \001(\t\022\022" +
+ "\n\nconfidence\030\004 \001(\002\022\023\n\013speaker_tag\030\005 \001(\0052" +
+ "\332\003\n\006Speech\022\226\001\n\tRecognize\022/.google.cloud." +
+ "speech.v1p1beta1.RecognizeRequest\0320.goog" +
+ "le.cloud.speech.v1p1beta1.RecognizeRespo" +
+ "nse\"&\202\323\344\223\002 \"\033/v1p1beta1/speech:recognize" +
+ ":\001*\022\244\001\n\024LongRunningRecognize\022:.google.cl" +
+ "oud.speech.v1p1beta1.LongRunningRecogniz" +
+ "eRequest\032\035.google.longrunning.Operation\"" +
+ "1\202\323\344\223\002+\"&/v1p1beta1/speech:longrunningre" +
+ "cognize:\001*\022\217\001\n\022StreamingRecognize\0228.goog" +
+ "le.cloud.speech.v1p1beta1.StreamingRecog" +
+ "nizeRequest\0329.google.cloud.speech.v1p1be" +
+ "ta1.StreamingRecognizeResponse\"\000(\0010\001Bz\n!" +
+ "com.google.cloud.speech.v1p1beta1B\013Speec" +
+ "hProtoP\001ZCgoogle.golang.org/genproto/goo" +
+ "gleapis/cloud/speech/v1p1beta1;speech\370\001\001" +
+ "b\006proto3"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() {
@@ -241,6 +243,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors(
com.google.longrunning.OperationsProto.getDescriptor(),
com.google.protobuf.AnyProto.getDescriptor(),
com.google.protobuf.DurationProto.getDescriptor(),
+ com.google.protobuf.EmptyProto.getDescriptor(),
com.google.protobuf.TimestampProto.getDescriptor(),
com.google.rpc.StatusProto.getDescriptor(),
}, assigner);
@@ -321,7 +324,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors(
internal_static_google_cloud_speech_v1p1beta1_StreamingRecognitionResult_fieldAccessorTable = new
com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_google_cloud_speech_v1p1beta1_StreamingRecognitionResult_descriptor,
- new java.lang.String[] { "Alternatives", "IsFinal", "Stability", "ChannelTag", "LanguageCode", });
+ new java.lang.String[] { "Alternatives", "IsFinal", "Stability", "ResultEndTime", "ChannelTag", "LanguageCode", });
internal_static_google_cloud_speech_v1p1beta1_SpeechRecognitionResult_descriptor =
getDescriptor().getMessageTypes().get(13);
internal_static_google_cloud_speech_v1p1beta1_SpeechRecognitionResult_fieldAccessorTable = new
@@ -349,6 +352,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors(
com.google.longrunning.OperationsProto.getDescriptor();
com.google.protobuf.AnyProto.getDescriptor();
com.google.protobuf.DurationProto.getDescriptor();
+ com.google.protobuf.EmptyProto.getDescriptor();
com.google.protobuf.TimestampProto.getDescriptor();
com.google.rpc.StatusProto.getDescriptor();
}
diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechRecognitionAlternative.java b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechRecognitionAlternative.java
index 2e53e4a1e11b..35690d14eb80 100644
--- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechRecognitionAlternative.java
+++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechRecognitionAlternative.java
@@ -171,7 +171,7 @@ public float getConfidence() {
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -183,7 +183,7 @@ public java.util.List
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -196,7 +196,7 @@ public java.util.List
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -208,7 +208,7 @@ public int getWordsCount() {
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -220,7 +220,7 @@ public com.google.cloud.speech.v1p1beta1.WordInfo getWords(int index) {
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -777,7 +777,7 @@ private void ensureWordsIsMutable() {
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -793,7 +793,7 @@ public java.util.List
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -809,7 +809,7 @@ public int getWordsCount() {
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -825,7 +825,7 @@ public com.google.cloud.speech.v1p1beta1.WordInfo getWords(int index) {
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -848,7 +848,7 @@ public Builder setWords(
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -868,7 +868,7 @@ public Builder setWords(
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -890,7 +890,7 @@ public Builder addWords(com.google.cloud.speech.v1p1beta1.WordInfo value) {
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -913,7 +913,7 @@ public Builder addWords(
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -933,7 +933,7 @@ public Builder addWords(
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -953,7 +953,7 @@ public Builder addWords(
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -974,7 +974,7 @@ public Builder addAllWords(
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -993,7 +993,7 @@ public Builder clearWords() {
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -1012,7 +1012,7 @@ public Builder removeWords(int index) {
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -1025,7 +1025,7 @@ public com.google.cloud.speech.v1p1beta1.WordInfo.Builder getWordsBuilder(
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -1041,7 +1041,7 @@ public com.google.cloud.speech.v1p1beta1.WordInfoOrBuilder getWordsOrBuilder(
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -1058,7 +1058,7 @@ public com.google.cloud.speech.v1p1beta1.WordInfoOrBuilder getWordsOrBuilder(
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -1071,7 +1071,7 @@ public com.google.cloud.speech.v1p1beta1.WordInfo.Builder addWordsBuilder() {
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -1085,7 +1085,7 @@ public com.google.cloud.speech.v1p1beta1.WordInfo.Builder addWordsBuilder(
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechRecognitionAlternativeOrBuilder.java b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechRecognitionAlternativeOrBuilder.java
index 97516e4ee893..0fad54139b00 100644
--- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechRecognitionAlternativeOrBuilder.java
+++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/SpeechRecognitionAlternativeOrBuilder.java
@@ -43,7 +43,7 @@ public interface SpeechRecognitionAlternativeOrBuilder extends
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -54,7 +54,7 @@ public interface SpeechRecognitionAlternativeOrBuilder extends
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -64,7 +64,7 @@ public interface SpeechRecognitionAlternativeOrBuilder extends
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -74,7 +74,7 @@ public interface SpeechRecognitionAlternativeOrBuilder extends
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
@@ -85,7 +85,7 @@ public interface SpeechRecognitionAlternativeOrBuilder extends
/**
*
* Output only. A list of word-specific information for each recognized word.
- * Note: When enable_speaker_diarization is true, you will see all the words
+ * Note: When `enable_speaker_diarization` is true, you will see all the words
* from the beginning of the audio.
*
*
diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognitionResult.java b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognitionResult.java
index e4e45bd2b3b6..b04b6a62bf1f 100644
--- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognitionResult.java
+++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognitionResult.java
@@ -71,6 +71,19 @@ private StreamingRecognitionResult(
stability_ = input.readFloat();
break;
}
+ case 34: {
+ com.google.protobuf.Duration.Builder subBuilder = null;
+ if (resultEndTime_ != null) {
+ subBuilder = resultEndTime_.toBuilder();
+ }
+ resultEndTime_ = input.readMessage(com.google.protobuf.Duration.parser(), extensionRegistry);
+ if (subBuilder != null) {
+ subBuilder.mergeFrom(resultEndTime_);
+ resultEndTime_ = subBuilder.buildPartial();
+ }
+
+ break;
+ }
case 40: {
channelTag_ = input.readInt32();
@@ -222,6 +235,42 @@ public float getStability() {
return stability_;
}
+ public static final int RESULT_END_TIME_FIELD_NUMBER = 4;
+ private com.google.protobuf.Duration resultEndTime_;
+ /**
+ * + * Output only. Time offset of the end of this result relative to the + * beginning of the audio. + *+ * + *
.google.protobuf.Duration result_end_time = 4;
+ */
+ public boolean hasResultEndTime() {
+ return resultEndTime_ != null;
+ }
+ /**
+ * + * Output only. Time offset of the end of this result relative to the + * beginning of the audio. + *+ * + *
.google.protobuf.Duration result_end_time = 4;
+ */
+ public com.google.protobuf.Duration getResultEndTime() {
+ return resultEndTime_ == null ? com.google.protobuf.Duration.getDefaultInstance() : resultEndTime_;
+ }
+ /**
+ * + * Output only. Time offset of the end of this result relative to the + * beginning of the audio. + *+ * + *
.google.protobuf.Duration result_end_time = 4;
+ */
+ public com.google.protobuf.DurationOrBuilder getResultEndTimeOrBuilder() {
+ return getResultEndTime();
+ }
+
public static final int CHANNEL_TAG_FIELD_NUMBER = 5;
private int channelTag_;
/**
@@ -308,6 +357,9 @@ public void writeTo(com.google.protobuf.CodedOutputStream output)
if (stability_ != 0F) {
output.writeFloat(3, stability_);
}
+ if (resultEndTime_ != null) {
+ output.writeMessage(4, getResultEndTime());
+ }
if (channelTag_ != 0) {
output.writeInt32(5, channelTag_);
}
@@ -335,6 +387,10 @@ public int getSerializedSize() {
size += com.google.protobuf.CodedOutputStream
.computeFloatSize(3, stability_);
}
+ if (resultEndTime_ != null) {
+ size += com.google.protobuf.CodedOutputStream
+ .computeMessageSize(4, getResultEndTime());
+ }
if (channelTag_ != 0) {
size += com.google.protobuf.CodedOutputStream
.computeInt32Size(5, channelTag_);
@@ -366,6 +422,11 @@ public boolean equals(final java.lang.Object obj) {
java.lang.Float.floatToIntBits(getStability())
== java.lang.Float.floatToIntBits(
other.getStability()));
+ result = result && (hasResultEndTime() == other.hasResultEndTime());
+ if (hasResultEndTime()) {
+ result = result && getResultEndTime()
+ .equals(other.getResultEndTime());
+ }
result = result && (getChannelTag()
== other.getChannelTag());
result = result && getLanguageCode()
@@ -391,6 +452,10 @@ public int hashCode() {
hash = (37 * hash) + STABILITY_FIELD_NUMBER;
hash = (53 * hash) + java.lang.Float.floatToIntBits(
getStability());
+ if (hasResultEndTime()) {
+ hash = (37 * hash) + RESULT_END_TIME_FIELD_NUMBER;
+ hash = (53 * hash) + getResultEndTime().hashCode();
+ }
hash = (37 * hash) + CHANNEL_TAG_FIELD_NUMBER;
hash = (53 * hash) + getChannelTag();
hash = (37 * hash) + LANGUAGE_CODE_FIELD_NUMBER;
@@ -544,6 +609,12 @@ public Builder clear() {
stability_ = 0F;
+ if (resultEndTimeBuilder_ == null) {
+ resultEndTime_ = null;
+ } else {
+ resultEndTime_ = null;
+ resultEndTimeBuilder_ = null;
+ }
channelTag_ = 0;
languageCode_ = "";
@@ -587,6 +658,11 @@ public com.google.cloud.speech.v1p1beta1.StreamingRecognitionResult buildPartial
}
result.isFinal_ = isFinal_;
result.stability_ = stability_;
+ if (resultEndTimeBuilder_ == null) {
+ result.resultEndTime_ = resultEndTime_;
+ } else {
+ result.resultEndTime_ = resultEndTimeBuilder_.build();
+ }
result.channelTag_ = channelTag_;
result.languageCode_ = languageCode_;
result.bitField0_ = to_bitField0_;
@@ -670,6 +746,9 @@ public Builder mergeFrom(com.google.cloud.speech.v1p1beta1.StreamingRecognitionR
if (other.getStability() != 0F) {
setStability(other.getStability());
}
+ if (other.hasResultEndTime()) {
+ mergeResultEndTime(other.getResultEndTime());
+ }
if (other.getChannelTag() != 0) {
setChannelTag(other.getChannelTag());
}
@@ -1173,6 +1252,168 @@ public Builder clearStability() {
return this;
}
+ private com.google.protobuf.Duration resultEndTime_ = null;
+ private com.google.protobuf.SingleFieldBuilderV3<
+ com.google.protobuf.Duration, com.google.protobuf.Duration.Builder, com.google.protobuf.DurationOrBuilder> resultEndTimeBuilder_;
+ /**
+ * + * Output only. Time offset of the end of this result relative to the + * beginning of the audio. + *+ * + *
.google.protobuf.Duration result_end_time = 4;
+ */
+ public boolean hasResultEndTime() {
+ return resultEndTimeBuilder_ != null || resultEndTime_ != null;
+ }
+ /**
+ * + * Output only. Time offset of the end of this result relative to the + * beginning of the audio. + *+ * + *
.google.protobuf.Duration result_end_time = 4;
+ */
+ public com.google.protobuf.Duration getResultEndTime() {
+ if (resultEndTimeBuilder_ == null) {
+ return resultEndTime_ == null ? com.google.protobuf.Duration.getDefaultInstance() : resultEndTime_;
+ } else {
+ return resultEndTimeBuilder_.getMessage();
+ }
+ }
+ /**
+ * + * Output only. Time offset of the end of this result relative to the + * beginning of the audio. + *+ * + *
.google.protobuf.Duration result_end_time = 4;
+ */
+ public Builder setResultEndTime(com.google.protobuf.Duration value) {
+ if (resultEndTimeBuilder_ == null) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+ resultEndTime_ = value;
+ onChanged();
+ } else {
+ resultEndTimeBuilder_.setMessage(value);
+ }
+
+ return this;
+ }
+ /**
+ * + * Output only. Time offset of the end of this result relative to the + * beginning of the audio. + *+ * + *
.google.protobuf.Duration result_end_time = 4;
+ */
+ public Builder setResultEndTime(
+ com.google.protobuf.Duration.Builder builderForValue) {
+ if (resultEndTimeBuilder_ == null) {
+ resultEndTime_ = builderForValue.build();
+ onChanged();
+ } else {
+ resultEndTimeBuilder_.setMessage(builderForValue.build());
+ }
+
+ return this;
+ }
+ /**
+ * + * Output only. Time offset of the end of this result relative to the + * beginning of the audio. + *+ * + *
.google.protobuf.Duration result_end_time = 4;
+ */
+ public Builder mergeResultEndTime(com.google.protobuf.Duration value) {
+ if (resultEndTimeBuilder_ == null) {
+ if (resultEndTime_ != null) {
+ resultEndTime_ =
+ com.google.protobuf.Duration.newBuilder(resultEndTime_).mergeFrom(value).buildPartial();
+ } else {
+ resultEndTime_ = value;
+ }
+ onChanged();
+ } else {
+ resultEndTimeBuilder_.mergeFrom(value);
+ }
+
+ return this;
+ }
+ /**
+ * + * Output only. Time offset of the end of this result relative to the + * beginning of the audio. + *+ * + *
.google.protobuf.Duration result_end_time = 4;
+ */
+ public Builder clearResultEndTime() {
+ if (resultEndTimeBuilder_ == null) {
+ resultEndTime_ = null;
+ onChanged();
+ } else {
+ resultEndTime_ = null;
+ resultEndTimeBuilder_ = null;
+ }
+
+ return this;
+ }
+ /**
+ * + * Output only. Time offset of the end of this result relative to the + * beginning of the audio. + *+ * + *
.google.protobuf.Duration result_end_time = 4;
+ */
+ public com.google.protobuf.Duration.Builder getResultEndTimeBuilder() {
+
+ onChanged();
+ return getResultEndTimeFieldBuilder().getBuilder();
+ }
+ /**
+ * + * Output only. Time offset of the end of this result relative to the + * beginning of the audio. + *+ * + *
.google.protobuf.Duration result_end_time = 4;
+ */
+ public com.google.protobuf.DurationOrBuilder getResultEndTimeOrBuilder() {
+ if (resultEndTimeBuilder_ != null) {
+ return resultEndTimeBuilder_.getMessageOrBuilder();
+ } else {
+ return resultEndTime_ == null ?
+ com.google.protobuf.Duration.getDefaultInstance() : resultEndTime_;
+ }
+ }
+ /**
+ * + * Output only. Time offset of the end of this result relative to the + * beginning of the audio. + *+ * + *
.google.protobuf.Duration result_end_time = 4;
+ */
+ private com.google.protobuf.SingleFieldBuilderV3<
+ com.google.protobuf.Duration, com.google.protobuf.Duration.Builder, com.google.protobuf.DurationOrBuilder>
+ getResultEndTimeFieldBuilder() {
+ if (resultEndTimeBuilder_ == null) {
+ resultEndTimeBuilder_ = new com.google.protobuf.SingleFieldBuilderV3<
+ com.google.protobuf.Duration, com.google.protobuf.Duration.Builder, com.google.protobuf.DurationOrBuilder>(
+ getResultEndTime(),
+ getParentForChildren(),
+ isClean());
+ resultEndTime_ = null;
+ }
+ return resultEndTimeBuilder_;
+ }
+
private int channelTag_ ;
/**
*
diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognitionResultOrBuilder.java b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognitionResultOrBuilder.java
index 445015b286b7..2f39dd41f60c 100644
--- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognitionResultOrBuilder.java
+++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognitionResultOrBuilder.java
@@ -92,6 +92,34 @@ com.google.cloud.speech.v1p1beta1.SpeechRecognitionAlternativeOrBuilder getAlter
*/
float getStability();
+ /**
+ *
+ * Output only. Time offset of the end of this result relative to the
+ * beginning of the audio.
+ *
+ *
+ * .google.protobuf.Duration result_end_time = 4;
+ */
+ boolean hasResultEndTime();
+ /**
+ *
+ * Output only. Time offset of the end of this result relative to the
+ * beginning of the audio.
+ *
+ *
+ * .google.protobuf.Duration result_end_time = 4;
+ */
+ com.google.protobuf.Duration getResultEndTime();
+ /**
+ *
+ * Output only. Time offset of the end of this result relative to the
+ * beginning of the audio.
+ *
+ *
+ * .google.protobuf.Duration result_end_time = 4;
+ */
+ com.google.protobuf.DurationOrBuilder getResultEndTimeOrBuilder();
+
/**
*
* For multi-channel audio, this is the channel number corresponding to the
diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognizeRequest.java b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognizeRequest.java
index 5fa3c577f017..6ad9f098c637 100644
--- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognizeRequest.java
+++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognizeRequest.java
@@ -193,7 +193,7 @@ public com.google.cloud.speech.v1p1beta1.StreamingRecognitionConfigOrBuilder get
* `audio_content` data. The audio bytes must be encoded as specified in
* `RecognitionConfig`. Note: as with all bytes fields, protobuffers use a
* pure binary representation (not base64). See
- * [audio limits](https://cloud.google.com/speech/limits#content).
+ * [content limits](/speech-to-text/quotas#content).
*
*
* bytes audio_content = 2;
@@ -782,7 +782,7 @@ public com.google.cloud.speech.v1p1beta1.StreamingRecognitionConfigOrBuilder get
* `audio_content` data. The audio bytes must be encoded as specified in
* `RecognitionConfig`. Note: as with all bytes fields, protobuffers use a
* pure binary representation (not base64). See
- * [audio limits](https://cloud.google.com/speech/limits#content).
+ * [content limits](/speech-to-text/quotas#content).
*
*
* bytes audio_content = 2;
@@ -802,7 +802,7 @@ public com.google.protobuf.ByteString getAudioContent() {
* `audio_content` data. The audio bytes must be encoded as specified in
* `RecognitionConfig`. Note: as with all bytes fields, protobuffers use a
* pure binary representation (not base64). See
- * [audio limits](https://cloud.google.com/speech/limits#content).
+ * [content limits](/speech-to-text/quotas#content).
*
*
* bytes audio_content = 2;
@@ -825,7 +825,7 @@ public Builder setAudioContent(com.google.protobuf.ByteString value) {
* `audio_content` data. The audio bytes must be encoded as specified in
* `RecognitionConfig`. Note: as with all bytes fields, protobuffers use a
* pure binary representation (not base64). See
- * [audio limits](https://cloud.google.com/speech/limits#content).
+ * [content limits](/speech-to-text/quotas#content).
*
*
* bytes audio_content = 2;
diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognizeRequestOrBuilder.java b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognizeRequestOrBuilder.java
index 5dee786892ce..f669b09d7397 100644
--- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognizeRequestOrBuilder.java
+++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/java/com/google/cloud/speech/v1p1beta1/StreamingRecognizeRequestOrBuilder.java
@@ -47,7 +47,7 @@ public interface StreamingRecognizeRequestOrBuilder extends
* `audio_content` data. The audio bytes must be encoded as specified in
* `RecognitionConfig`. Note: as with all bytes fields, protobuffers use a
* pure binary representation (not base64). See
- * [audio limits](https://cloud.google.com/speech/limits#content).
+ * [content limits](/speech-to-text/quotas#content).
*
*
* bytes audio_content = 2;
diff --git a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/proto/google/cloud/speech/v1p1beta1/cloud_speech.proto b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/proto/google/cloud/speech/v1p1beta1/cloud_speech.proto
index 208f53633e02..53283d3b7d33 100644
--- a/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/proto/google/cloud/speech/v1p1beta1/cloud_speech.proto
+++ b/google-api-grpc/proto-google-cloud-speech-v1p1beta1/src/main/proto/google/cloud/speech/v1p1beta1/cloud_speech.proto
@@ -1,4 +1,4 @@
-// Copyright 2018 Google Inc.
+// Copyright 2018 Google LLC.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
@@ -11,6 +11,7 @@
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
+//
syntax = "proto3";
@@ -20,6 +21,7 @@ import "google/api/annotations.proto";
import "google/longrunning/operations.proto";
import "google/protobuf/any.proto";
import "google/protobuf/duration.proto";
+import "google/protobuf/empty.proto";
import "google/protobuf/timestamp.proto";
import "google/rpc/status.proto";
@@ -54,7 +56,8 @@ service Speech {
// Performs bidirectional streaming speech recognition: receive results while
// sending audio. This method is only available via the gRPC API (not REST).
- rpc StreamingRecognize(stream StreamingRecognizeRequest) returns (stream StreamingRecognizeResponse);
+ rpc StreamingRecognize(stream StreamingRecognizeRequest) returns (stream StreamingRecognizeResponse) {
+ }
}
// The top-level message sent by the client for the `Recognize` method.
@@ -98,7 +101,7 @@ message StreamingRecognizeRequest {
// `audio_content` data. The audio bytes must be encoded as specified in
// `RecognitionConfig`. Note: as with all bytes fields, protobuffers use a
// pure binary representation (not base64). See
- // [audio limits](https://cloud.google.com/speech/limits#content).
+ // [content limits](/speech-to-text/quotas#content).
bytes audio_content = 2;
}
}
@@ -218,36 +221,36 @@ message RecognitionConfig {
// Valid values for OGG_OPUS are '1'-'254'.
// Valid value for MULAW, AMR, AMR_WB and SPEEX_WITH_HEADER_BYTE is only `1`.
// If `0` or omitted, defaults to one channel (mono).
- // NOTE: We only recognize the first channel by default.
+ // Note: We only recognize the first channel by default.
// To perform independent recognition on each channel set
- // enable_separate_recognition_per_channel to 'true'.
+ // `enable_separate_recognition_per_channel` to 'true'.
int32 audio_channel_count = 7;
- // This needs to be set to ‘true’ explicitly and audio_channel_count > 1
+ // This needs to be set to ‘true’ explicitly and `audio_channel_count` > 1
// to get each channel recognized separately. The recognition result will
- // contain a channel_tag field to state which channel that result belongs to.
- // If this is not ‘true’, we will only recognize the first channel.
- // NOTE: The request is also billed cumulatively for all channels recognized:
- // (audio_channel_count times the audio length)
+ // contain a `channel_tag` field to state which channel that result belongs
+ // to. If this is not true, we will only recognize the first channel. The
+ // request is billed cumulatively for all channels recognized:
+ // `audio_channel_count` multiplied by the length of the audio.
bool enable_separate_recognition_per_channel = 12;
// *Required* The language of the supplied audio as a
// [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
// Example: "en-US".
- // See [Language Support](https://cloud.google.com/speech/docs/languages)
+ // See [Language Support](/speech-to-text/docs/languages)
// for a list of the currently supported language codes.
string language_code = 3;
// *Optional* A list of up to 3 additional
// [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tags,
// listing possible alternative languages of the supplied audio.
- // See [Language Support](https://cloud.google.com/speech/docs/languages)
+ // See [Language Support](/speech-to-text/docs/languages)
// for a list of the currently supported language codes.
// If alternative languages are listed, recognition result will contain
// recognition in the most likely language detected including the main
// language_code. The recognition result will include the language tag
// of the language detected in the audio.
- // NOTE: This feature is only supported for Voice Command and Voice Search
+ // Note: This feature is only supported for Voice Command and Voice Search
// use cases and performance may vary for other use cases (e.g., phone call
// transcription).
repeated string alternative_language_codes = 18;
@@ -266,7 +269,9 @@ message RecognitionConfig {
// won't be filtered out.
bool profanity_filter = 5;
- // *Optional* A means to provide context to assist the speech recognition.
+ // *Optional* array of [SpeechContext][google.cloud.speech.v1p1beta1.SpeechContext].
+ // A means to provide context to assist the speech recognition. For more
+ // information, see [Phrase Hints](/speech-to-text/docs/basics#phrase-hints).
repeated SpeechContext speech_contexts = 6;
// *Optional* If `true`, the top result includes a list of words and
@@ -284,18 +289,20 @@ message RecognitionConfig {
// This feature is only available in select languages. Setting this for
// requests in other languages has no effect at all.
// The default 'false' value does not add punctuation to result hypotheses.
- // NOTE: "This is currently offered as an experimental service, complimentary
+ // Note: This is currently offered as an experimental service, complimentary
// to all users. In the future this may be exclusively available as a
- // premium feature."
+ // premium feature.
bool enable_automatic_punctuation = 11;
// *Optional* If 'true', enables speaker detection for each recognized word in
// the top alternative of the recognition result using a speaker_tag provided
// in the WordInfo.
// Note: When this is true, we send all the words from the beginning of the
- // audio for the top alternative in every consecutive responses.
+ // audio for the top alternative in every consecutive STREAMING responses.
// This is done in order to improve our speaker tags as our models learn to
// identify the speakers in the conversation over time.
+ // For non-streaming requests, the diarization results will be provided only
+ // in the top alternative of the FINAL SpeechRecognitionResult.
bool enable_speaker_diarization = 16;
// *Optional*
@@ -342,14 +349,18 @@ message RecognitionConfig {
string model = 13;
// *Optional* Set to true to use an enhanced model for speech recognition.
- // You must also set the `model` field to a valid, enhanced model. If
- // `use_enhanced` is set to true and the `model` field is not set, then
- // `use_enhanced` is ignored. If `use_enhanced` is true and an enhanced
- // version of the specified model does not exist, then the speech is
- // recognized using the standard version of the specified model.
+ // If `use_enhanced` is set to true and the `model` field is not set, then
+ // an appropriate enhanced model is chosen if:
+ // 1. project is eligible for requesting enhanced models
+ // 2. an enhanced model exists for the audio
+ //
+ // If `use_enhanced` is true and an enhanced version of the specified model
+ // does not exist, then the speech is recognized using the standard version
+ // of the specified model.
//
- // Enhanced speech models require that you opt-in to the audio logging using
- // instructions in the [alpha documentation](/speech/data-sharing). If you set
+ // Enhanced speech models require that you opt-in to data logging using
+ // instructions in the
+ // [documentation](/speech-to-text/docs/enable-data-logging). If you set
// `use_enhanced` to true and you have not enabled audio logging, then you
// will receive an error.
bool use_enhanced = 14;
@@ -494,14 +505,14 @@ message SpeechContext {
// to improve the accuracy for specific words and phrases, for example, if
// specific commands are typically spoken by the user. This can also be used
// to add additional words to the vocabulary of the recognizer. See
- // [usage limits](https://cloud.google.com/speech/limits#content).
+ // [usage limits](/speech-to-text/quotas#content).
repeated string phrases = 1;
}
// Contains audio data in the encoding specified in the `RecognitionConfig`.
// Either `content` or `uri` must be supplied. Supplying both or neither
// returns [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]. See
-// [audio limits](https://cloud.google.com/speech/limits#content).
+// [content limits](/speech-to-text/quotas#content).
message RecognitionAudio {
// The audio source, which is either inline content or a Google Cloud
// Storage uri.
@@ -512,7 +523,8 @@ message RecognitionAudio {
bytes content = 1;
// URI that points to a file that contains audio data bytes as specified in
- // `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
+ // `RecognitionConfig`. The file must not be compressed (for example, gzip).
+ // Currently, only Google Cloud Storage URIs are
// supported, which must be specified in the following format:
// `gs://bucket_name/object_name` (other URI formats return
// [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
@@ -658,6 +670,10 @@ message StreamingRecognitionResult {
// The default of 0.0 is a sentinel value indicating `stability` was not set.
float stability = 3;
+ // Output only. Time offset of the end of this result relative to the
+ // beginning of the audio.
+ google.protobuf.Duration result_end_time = 4;
+
// For multi-channel audio, this is the channel number corresponding to the
// recognized result for the audio from that channel.
// For audio_channel_count = N, its output values can range from '1' to 'N'.
@@ -705,7 +721,7 @@ message SpeechRecognitionAlternative {
float confidence = 2;
// Output only. A list of word-specific information for each recognized word.
- // Note: When enable_speaker_diarization is true, you will see all the words
+ // Note: When `enable_speaker_diarization` is true, you will see all the words
// from the beginning of the audio.
repeated WordInfo words = 3;
}
@@ -746,5 +762,4 @@ message WordInfo {
// speaker_tag is set if enable_speaker_diarization = 'true' and only in the
// top alternative.
int32 speaker_tag = 5;
-
}