diff --git a/google-api-grpc/proto-google-cloud-speech-v1/src/main/java/com/google/cloud/speech/v1/RecognitionConfig.java b/google-api-grpc/proto-google-cloud-speech-v1/src/main/java/com/google/cloud/speech/v1/RecognitionConfig.java index 12f86d28f193..6904737bf13a 100644 --- a/google-api-grpc/proto-google-cloud-speech-v1/src/main/java/com/google/cloud/speech/v1/RecognitionConfig.java +++ b/google-api-grpc/proto-google-cloud-speech-v1/src/main/java/com/google/cloud/speech/v1/RecognitionConfig.java @@ -26,6 +26,7 @@ private RecognitionConfig(com.google.protobuf.GeneratedMessageV3.Builder buil private RecognitionConfig() { encoding_ = 0; sampleRateHertz_ = 0; + audioChannelCount_ = 0; enableSeparateRecognitionPerChannel_ = false; languageCode_ = ""; maxAlternatives_ = 0; @@ -92,16 +93,21 @@ private RecognitionConfig( } case 50: { - if (!((mutable_bitField0_ & 0x00000040) == 0x00000040)) { + if (!((mutable_bitField0_ & 0x00000080) == 0x00000080)) { speechContexts_ = new java.util.ArrayList(); - mutable_bitField0_ |= 0x00000040; + mutable_bitField0_ |= 0x00000080; } speechContexts_.add( input.readMessage( com.google.cloud.speech.v1.SpeechContext.parser(), extensionRegistry)); break; } + case 56: + { + audioChannelCount_ = input.readInt32(); + break; + } case 64: { enableWordTimeOffsets_ = input.readBool(); @@ -143,7 +149,7 @@ private RecognitionConfig( } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException(e).setUnfinishedMessage(this); } finally { - if (((mutable_bitField0_ & 0x00000040) == 0x00000040)) { + if (((mutable_bitField0_ & 0x00000080) == 0x00000080)) { speechContexts_ = java.util.Collections.unmodifiableList(speechContexts_); } this.unknownFields = unknownFields.build(); @@ -533,6 +539,29 @@ public int getSampleRateHertz() { return sampleRateHertz_; } + public static final int AUDIO_CHANNEL_COUNT_FIELD_NUMBER = 7; + private int audioChannelCount_; + /** + * + * + *
+   * *Optional* The number of channels in the input audio data.
+   * ONLY set this for MULTI-CHANNEL recognition.
+   * Valid values for LINEAR16 and FLAC are `1`-`8`.
+   * Valid values for OGG_OPUS are '1'-'254'.
+   * Valid value for MULAW, AMR, AMR_WB and SPEEX_WITH_HEADER_BYTE is only `1`.
+   * If `0` or omitted, defaults to one channel (mono).
+   * Note: We only recognize the first channel by default.
+   * To perform independent recognition on each channel set
+   * `enable_separate_recognition_per_channel` to 'true'.
+   * 
+ * + * int32 audio_channel_count = 7; + */ + public int getAudioChannelCount() { + return audioChannelCount_; + } + public static final int ENABLE_SEPARATE_RECOGNITION_PER_CHANNEL_FIELD_NUMBER = 12; private boolean enableSeparateRecognitionPerChannel_; /** @@ -921,6 +950,9 @@ public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io for (int i = 0; i < speechContexts_.size(); i++) { output.writeMessage(6, speechContexts_.get(i)); } + if (audioChannelCount_ != 0) { + output.writeInt32(7, audioChannelCount_); + } if (enableWordTimeOffsets_ != false) { output.writeBool(8, enableWordTimeOffsets_); } @@ -965,6 +997,9 @@ public int getSerializedSize() { for (int i = 0; i < speechContexts_.size(); i++) { size += com.google.protobuf.CodedOutputStream.computeMessageSize(6, speechContexts_.get(i)); } + if (audioChannelCount_ != 0) { + size += com.google.protobuf.CodedOutputStream.computeInt32Size(7, audioChannelCount_); + } if (enableWordTimeOffsets_ != false) { size += com.google.protobuf.CodedOutputStream.computeBoolSize(8, enableWordTimeOffsets_); } @@ -1002,6 +1037,7 @@ public boolean equals(final java.lang.Object obj) { boolean result = true; result = result && encoding_ == other.encoding_; result = result && (getSampleRateHertz() == other.getSampleRateHertz()); + result = result && (getAudioChannelCount() == other.getAudioChannelCount()); result = result && (getEnableSeparateRecognitionPerChannel() @@ -1029,6 +1065,8 @@ public int hashCode() { hash = (53 * hash) + encoding_; hash = (37 * hash) + SAMPLE_RATE_HERTZ_FIELD_NUMBER; hash = (53 * hash) + getSampleRateHertz(); + hash = (37 * hash) + AUDIO_CHANNEL_COUNT_FIELD_NUMBER; + hash = (53 * hash) + getAudioChannelCount(); hash = (37 * hash) + ENABLE_SEPARATE_RECOGNITION_PER_CHANNEL_FIELD_NUMBER; hash = (53 * hash) @@ -1203,6 +1241,8 @@ public Builder clear() { sampleRateHertz_ = 0; + audioChannelCount_ = 0; + enableSeparateRecognitionPerChannel_ = false; languageCode_ = ""; @@ -1213,7 +1253,7 @@ public Builder clear() { if (speechContextsBuilder_ == null) { speechContexts_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000040); + bitField0_ = (bitField0_ & ~0x00000080); } else { speechContextsBuilder_.clear(); } @@ -1256,14 +1296,15 @@ public com.google.cloud.speech.v1.RecognitionConfig buildPartial() { int to_bitField0_ = 0; result.encoding_ = encoding_; result.sampleRateHertz_ = sampleRateHertz_; + result.audioChannelCount_ = audioChannelCount_; result.enableSeparateRecognitionPerChannel_ = enableSeparateRecognitionPerChannel_; result.languageCode_ = languageCode_; result.maxAlternatives_ = maxAlternatives_; result.profanityFilter_ = profanityFilter_; if (speechContextsBuilder_ == null) { - if (((bitField0_ & 0x00000040) == 0x00000040)) { + if (((bitField0_ & 0x00000080) == 0x00000080)) { speechContexts_ = java.util.Collections.unmodifiableList(speechContexts_); - bitField0_ = (bitField0_ & ~0x00000040); + bitField0_ = (bitField0_ & ~0x00000080); } result.speechContexts_ = speechContexts_; } else { @@ -1329,6 +1370,9 @@ public Builder mergeFrom(com.google.cloud.speech.v1.RecognitionConfig other) { if (other.getSampleRateHertz() != 0) { setSampleRateHertz(other.getSampleRateHertz()); } + if (other.getAudioChannelCount() != 0) { + setAudioChannelCount(other.getAudioChannelCount()); + } if (other.getEnableSeparateRecognitionPerChannel() != false) { setEnableSeparateRecognitionPerChannel(other.getEnableSeparateRecognitionPerChannel()); } @@ -1346,7 +1390,7 @@ public Builder mergeFrom(com.google.cloud.speech.v1.RecognitionConfig other) { if (!other.speechContexts_.isEmpty()) { if (speechContexts_.isEmpty()) { speechContexts_ = other.speechContexts_; - bitField0_ = (bitField0_ & ~0x00000040); + bitField0_ = (bitField0_ & ~0x00000080); } else { ensureSpeechContextsIsMutable(); speechContexts_.addAll(other.speechContexts_); @@ -1359,7 +1403,7 @@ public Builder mergeFrom(com.google.cloud.speech.v1.RecognitionConfig other) { speechContextsBuilder_.dispose(); speechContextsBuilder_ = null; speechContexts_ = other.speechContexts_; - bitField0_ = (bitField0_ & ~0x00000040); + bitField0_ = (bitField0_ & ~0x00000080); speechContextsBuilder_ = com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getSpeechContextsFieldBuilder() @@ -1563,6 +1607,74 @@ public Builder clearSampleRateHertz() { return this; } + private int audioChannelCount_; + /** + * + * + *
+     * *Optional* The number of channels in the input audio data.
+     * ONLY set this for MULTI-CHANNEL recognition.
+     * Valid values for LINEAR16 and FLAC are `1`-`8`.
+     * Valid values for OGG_OPUS are '1'-'254'.
+     * Valid value for MULAW, AMR, AMR_WB and SPEEX_WITH_HEADER_BYTE is only `1`.
+     * If `0` or omitted, defaults to one channel (mono).
+     * Note: We only recognize the first channel by default.
+     * To perform independent recognition on each channel set
+     * `enable_separate_recognition_per_channel` to 'true'.
+     * 
+ * + * int32 audio_channel_count = 7; + */ + public int getAudioChannelCount() { + return audioChannelCount_; + } + /** + * + * + *
+     * *Optional* The number of channels in the input audio data.
+     * ONLY set this for MULTI-CHANNEL recognition.
+     * Valid values for LINEAR16 and FLAC are `1`-`8`.
+     * Valid values for OGG_OPUS are '1'-'254'.
+     * Valid value for MULAW, AMR, AMR_WB and SPEEX_WITH_HEADER_BYTE is only `1`.
+     * If `0` or omitted, defaults to one channel (mono).
+     * Note: We only recognize the first channel by default.
+     * To perform independent recognition on each channel set
+     * `enable_separate_recognition_per_channel` to 'true'.
+     * 
+ * + * int32 audio_channel_count = 7; + */ + public Builder setAudioChannelCount(int value) { + + audioChannelCount_ = value; + onChanged(); + return this; + } + /** + * + * + *
+     * *Optional* The number of channels in the input audio data.
+     * ONLY set this for MULTI-CHANNEL recognition.
+     * Valid values for LINEAR16 and FLAC are `1`-`8`.
+     * Valid values for OGG_OPUS are '1'-'254'.
+     * Valid value for MULAW, AMR, AMR_WB and SPEEX_WITH_HEADER_BYTE is only `1`.
+     * If `0` or omitted, defaults to one channel (mono).
+     * Note: We only recognize the first channel by default.
+     * To perform independent recognition on each channel set
+     * `enable_separate_recognition_per_channel` to 'true'.
+     * 
+ * + * int32 audio_channel_count = 7; + */ + public Builder clearAudioChannelCount() { + + audioChannelCount_ = 0; + onChanged(); + return this; + } + private boolean enableSeparateRecognitionPerChannel_; /** * @@ -1852,10 +1964,10 @@ public Builder clearProfanityFilter() { java.util.Collections.emptyList(); private void ensureSpeechContextsIsMutable() { - if (!((bitField0_ & 0x00000040) == 0x00000040)) { + if (!((bitField0_ & 0x00000080) == 0x00000080)) { speechContexts_ = new java.util.ArrayList(speechContexts_); - bitField0_ |= 0x00000040; + bitField0_ |= 0x00000080; } } @@ -2093,7 +2205,7 @@ public Builder addAllSpeechContexts( public Builder clearSpeechContexts() { if (speechContextsBuilder_ == null) { speechContexts_ = java.util.Collections.emptyList(); - bitField0_ = (bitField0_ & ~0x00000040); + bitField0_ = (bitField0_ & ~0x00000080); onChanged(); } else { speechContextsBuilder_.clear(); @@ -2230,7 +2342,7 @@ public com.google.cloud.speech.v1.SpeechContext.Builder addSpeechContextsBuilder com.google.cloud.speech.v1.SpeechContext.Builder, com.google.cloud.speech.v1.SpeechContextOrBuilder>( speechContexts_, - ((bitField0_ & 0x00000040) == 0x00000040), + ((bitField0_ & 0x00000080) == 0x00000080), getParentForChildren(), isClean()); speechContexts_ = null; diff --git a/google-api-grpc/proto-google-cloud-speech-v1/src/main/java/com/google/cloud/speech/v1/RecognitionConfigOrBuilder.java b/google-api-grpc/proto-google-cloud-speech-v1/src/main/java/com/google/cloud/speech/v1/RecognitionConfigOrBuilder.java index 52da3c26b3dd..43532b9f87e9 100644 --- a/google-api-grpc/proto-google-cloud-speech-v1/src/main/java/com/google/cloud/speech/v1/RecognitionConfigOrBuilder.java +++ b/google-api-grpc/proto-google-cloud-speech-v1/src/main/java/com/google/cloud/speech/v1/RecognitionConfigOrBuilder.java @@ -50,6 +50,25 @@ public interface RecognitionConfigOrBuilder */ int getSampleRateHertz(); + /** + * + * + *
+   * *Optional* The number of channels in the input audio data.
+   * ONLY set this for MULTI-CHANNEL recognition.
+   * Valid values for LINEAR16 and FLAC are `1`-`8`.
+   * Valid values for OGG_OPUS are '1'-'254'.
+   * Valid value for MULAW, AMR, AMR_WB and SPEEX_WITH_HEADER_BYTE is only `1`.
+   * If `0` or omitted, defaults to one channel (mono).
+   * Note: We only recognize the first channel by default.
+   * To perform independent recognition on each channel set
+   * `enable_separate_recognition_per_channel` to 'true'.
+   * 
+ * + * int32 audio_channel_count = 7; + */ + int getAudioChannelCount(); + /** * * diff --git a/google-api-grpc/proto-google-cloud-speech-v1/src/main/java/com/google/cloud/speech/v1/SpeechProto.java b/google-api-grpc/proto-google-cloud-speech-v1/src/main/java/com/google/cloud/speech/v1/SpeechProto.java index 3dac3ed10b25..73fe98c9ee00 100644 --- a/google-api-grpc/proto-google-cloud-speech-v1/src/main/java/com/google/cloud/speech/v1/SpeechProto.java +++ b/google-api-grpc/proto-google-cloud-speech-v1/src/main/java/com/google/cloud/speech/v1/SpeechProto.java @@ -102,67 +102,68 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "t\"\212\001\n\032StreamingRecognitionConfig\0229\n\006conf" + "ig\030\001 \001(\0132).google.cloud.speech.v1.Recogn" + "itionConfig\022\030\n\020single_utterance\030\002 \001(\010\022\027\n" - + "\017interim_results\030\003 \001(\010\"\260\004\n\021RecognitionCo" + + "\017interim_results\030\003 \001(\010\"\315\004\n\021RecognitionCo" + "nfig\022I\n\010encoding\030\001 \001(\01627.google.cloud.sp" + "eech.v1.RecognitionConfig.AudioEncoding\022" - + "\031\n\021sample_rate_hertz\030\002 \001(\005\022/\n\'enable_sep" - + "arate_recognition_per_channel\030\014 \001(\010\022\025\n\rl" - + "anguage_code\030\003 \001(\t\022\030\n\020max_alternatives\030\004" - + " \001(\005\022\030\n\020profanity_filter\030\005 \001(\010\022>\n\017speech" - + "_contexts\030\006 \003(\0132%.google.cloud.speech.v1" - + ".SpeechContext\022 \n\030enable_word_time_offse" - + "ts\030\010 \001(\010\022$\n\034enable_automatic_punctuation" - + "\030\013 \001(\010\022\r\n\005model\030\r \001(\t\022\024\n\014use_enhanced\030\016 " - + "\001(\010\"\213\001\n\rAudioEncoding\022\030\n\024ENCODING_UNSPEC" - + "IFIED\020\000\022\014\n\010LINEAR16\020\001\022\010\n\004FLAC\020\002\022\t\n\005MULAW" - + "\020\003\022\007\n\003AMR\020\004\022\n\n\006AMR_WB\020\005\022\014\n\010OGG_OPUS\020\006\022\032\n" - + "\026SPEEX_WITH_HEADER_BYTE\020\007\" \n\rSpeechConte" - + "xt\022\017\n\007phrases\030\001 \003(\t\"D\n\020RecognitionAudio\022" - + "\021\n\007content\030\001 \001(\014H\000\022\r\n\003uri\030\002 \001(\tH\000B\016\n\014aud" - + "io_source\"U\n\021RecognizeResponse\022@\n\007result" - + "s\030\002 \003(\0132/.google.cloud.speech.v1.SpeechR" - + "ecognitionResult\"`\n\034LongRunningRecognize" - + "Response\022@\n\007results\030\002 \003(\0132/.google.cloud" - + ".speech.v1.SpeechRecognitionResult\"\236\001\n\034L" - + "ongRunningRecognizeMetadata\022\030\n\020progress_" - + "percent\030\001 \001(\005\022.\n\nstart_time\030\002 \001(\0132\032.goog" - + "le.protobuf.Timestamp\0224\n\020last_update_tim" - + "e\030\003 \001(\0132\032.google.protobuf.Timestamp\"\261\002\n\032" - + "StreamingRecognizeResponse\022!\n\005error\030\001 \001(" - + "\0132\022.google.rpc.Status\022C\n\007results\030\002 \003(\01322" - + ".google.cloud.speech.v1.StreamingRecogni" - + "tionResult\022]\n\021speech_event_type\030\004 \001(\0162B." - + "google.cloud.speech.v1.StreamingRecogniz" - + "eResponse.SpeechEventType\"L\n\017SpeechEvent" - + "Type\022\034\n\030SPEECH_EVENT_UNSPECIFIED\020\000\022\033\n\027EN" - + "D_OF_SINGLE_UTTERANCE\020\001\"\242\001\n\032StreamingRec" - + "ognitionResult\022J\n\014alternatives\030\001 \003(\01324.g" - + "oogle.cloud.speech.v1.SpeechRecognitionA" - + "lternative\022\020\n\010is_final\030\002 \001(\010\022\021\n\tstabilit" - + "y\030\003 \001(\002\022\023\n\013channel_tag\030\005 \001(\005\"z\n\027SpeechRe" - + "cognitionResult\022J\n\014alternatives\030\001 \003(\01324." - + "google.cloud.speech.v1.SpeechRecognition" - + "Alternative\022\023\n\013channel_tag\030\002 \001(\005\"w\n\034Spee" - + "chRecognitionAlternative\022\022\n\ntranscript\030\001" - + " \001(\t\022\022\n\nconfidence\030\002 \001(\002\022/\n\005words\030\003 \003(\0132" - + " .google.cloud.speech.v1.WordInfo\"t\n\010Wor" - + "dInfo\022-\n\nstart_time\030\001 \001(\0132\031.google.proto" - + "buf.Duration\022+\n\010end_time\030\002 \001(\0132\031.google." - + "protobuf.Duration\022\014\n\004word\030\003 \001(\t2\251\003\n\006Spee" - + "ch\022\201\001\n\tRecognize\022(.google.cloud.speech.v" - + "1.RecognizeRequest\032).google.cloud.speech" - + ".v1.RecognizeResponse\"\037\202\323\344\223\002\031\"\024/v1/speec" - + "h:recognize:\001*\022\226\001\n\024LongRunningRecognize\022" - + "3.google.cloud.speech.v1.LongRunningReco" - + "gnizeRequest\032\035.google.longrunning.Operat" - + "ion\"*\202\323\344\223\002$\"\037/v1/speech:longrunningrecog" - + "nize:\001*\022\201\001\n\022StreamingRecognize\0221.google." - + "cloud.speech.v1.StreamingRecognizeReques" - + "t\0322.google.cloud.speech.v1.StreamingReco" - + "gnizeResponse\"\000(\0010\001Bl\n\032com.google.cloud." - + "speech.v1B\013SpeechProtoP\001Z\n\017speech_contexts\030\006" + + " \003(\0132%.google.cloud.speech.v1.SpeechCont" + + "ext\022 \n\030enable_word_time_offsets\030\010 \001(\010\022$\n" + + "\034enable_automatic_punctuation\030\013 \001(\010\022\r\n\005m" + + "odel\030\r \001(\t\022\024\n\014use_enhanced\030\016 \001(\010\"\213\001\n\rAud" + + "ioEncoding\022\030\n\024ENCODING_UNSPECIFIED\020\000\022\014\n\010" + + "LINEAR16\020\001\022\010\n\004FLAC\020\002\022\t\n\005MULAW\020\003\022\007\n\003AMR\020\004" + + "\022\n\n\006AMR_WB\020\005\022\014\n\010OGG_OPUS\020\006\022\032\n\026SPEEX_WITH" + + "_HEADER_BYTE\020\007\" \n\rSpeechContext\022\017\n\007phras" + + "es\030\001 \003(\t\"D\n\020RecognitionAudio\022\021\n\007content\030" + + "\001 \001(\014H\000\022\r\n\003uri\030\002 \001(\tH\000B\016\n\014audio_source\"U" + + "\n\021RecognizeResponse\022@\n\007results\030\002 \003(\0132/.g" + + "oogle.cloud.speech.v1.SpeechRecognitionR" + + "esult\"`\n\034LongRunningRecognizeResponse\022@\n" + + "\007results\030\002 \003(\0132/.google.cloud.speech.v1." + + "SpeechRecognitionResult\"\236\001\n\034LongRunningR" + + "ecognizeMetadata\022\030\n\020progress_percent\030\001 \001" + + "(\005\022.\n\nstart_time\030\002 \001(\0132\032.google.protobuf" + + ".Timestamp\0224\n\020last_update_time\030\003 \001(\0132\032.g" + + "oogle.protobuf.Timestamp\"\261\002\n\032StreamingRe" + + "cognizeResponse\022!\n\005error\030\001 \001(\0132\022.google." + + "rpc.Status\022C\n\007results\030\002 \003(\01322.google.clo" + + "ud.speech.v1.StreamingRecognitionResult\022" + + "]\n\021speech_event_type\030\004 \001(\0162B.google.clou" + + "d.speech.v1.StreamingRecognizeResponse.S" + + "peechEventType\"L\n\017SpeechEventType\022\034\n\030SPE" + + "ECH_EVENT_UNSPECIFIED\020\000\022\033\n\027END_OF_SINGLE" + + "_UTTERANCE\020\001\"\242\001\n\032StreamingRecognitionRes" + + "ult\022J\n\014alternatives\030\001 \003(\01324.google.cloud" + + ".speech.v1.SpeechRecognitionAlternative\022" + + "\020\n\010is_final\030\002 \001(\010\022\021\n\tstability\030\003 \001(\002\022\023\n\013" + + "channel_tag\030\005 \001(\005\"z\n\027SpeechRecognitionRe" + + "sult\022J\n\014alternatives\030\001 \003(\01324.google.clou" + + "d.speech.v1.SpeechRecognitionAlternative" + + "\022\023\n\013channel_tag\030\002 \001(\005\"w\n\034SpeechRecogniti" + + "onAlternative\022\022\n\ntranscript\030\001 \001(\t\022\022\n\ncon" + + "fidence\030\002 \001(\002\022/\n\005words\030\003 \003(\0132 .google.cl" + + "oud.speech.v1.WordInfo\"t\n\010WordInfo\022-\n\nst" + + "art_time\030\001 \001(\0132\031.google.protobuf.Duratio" + + "n\022+\n\010end_time\030\002 \001(\0132\031.google.protobuf.Du" + + "ration\022\014\n\004word\030\003 \001(\t2\251\003\n\006Speech\022\201\001\n\tReco" + + "gnize\022(.google.cloud.speech.v1.Recognize" + + "Request\032).google.cloud.speech.v1.Recogni" + + "zeResponse\"\037\202\323\344\223\002\031\"\024/v1/speech:recognize" + + ":\001*\022\226\001\n\024LongRunningRecognize\0223.google.cl" + + "oud.speech.v1.LongRunningRecognizeReques" + + "t\032\035.google.longrunning.Operation\"*\202\323\344\223\002$" + + "\"\037/v1/speech:longrunningrecognize:\001*\022\201\001\n" + + "\022StreamingRecognize\0221.google.cloud.speec" + + "h.v1.StreamingRecognizeRequest\0322.google." + + "cloud.speech.v1.StreamingRecognizeRespon" + + "se\"\000(\0010\001Bl\n\032com.google.cloud.speech.v1B\013" + + "SpeechProtoP\001Z 1 // to get each channel recognized separately. The recognition result will // contain a `channel_tag` field to state which channel that result belongs diff --git a/google-cloud-clients/google-cloud-speech/synth.metadata b/google-cloud-clients/google-cloud-speech/synth.metadata index f16c7066f21e..77ee6a1fa435 100644 --- a/google-cloud-clients/google-cloud-speech/synth.metadata +++ b/google-cloud-clients/google-cloud-speech/synth.metadata @@ -1,19 +1,19 @@ { - "updateTime": "2019-01-17T08:50:59.718229Z", + "updateTime": "2019-02-01T08:47:48.673490Z", "sources": [ { "generator": { "name": "artman", - "version": "0.16.6", - "dockerImage": "googleapis/artman@sha256:12722f2ca3fbc3b53cc6aa5f0e569d7d221b46bd876a2136497089dec5e3634e" + "version": "0.16.7", + "dockerImage": "googleapis/artman@sha256:d6c8ced606eb49973ca95d2af7c55a681acc042db0f87d135968349e7bf6dd80" } }, { "git": { "name": "googleapis", "remote": "https://github.com/googleapis/googleapis.git", - "sha": "0ac60e21a1aa86c07c1836865b35308ba8178b05", - "internalRef": "229626798" + "sha": "acb5253cd11cd43cab93eb153d6e48ba0fa5303d", + "internalRef": "231786007" } } ],