From 841930a680ebe370c9fc6ae824465f6bc51a5a46 Mon Sep 17 00:00:00 2001 From: yoshi-automation Date: Tue, 30 Apr 2019 00:42:19 -0700 Subject: [PATCH] Regenerate dialogflow client --- .../cloud/dialogflow/v2/AudioConfigProto.java | 84 ++++--- .../cloud/dialogflow/v2/AudioEncoding.java | 4 +- .../cloud/dialogflow/v2/InputAudioConfig.java | 147 +++++++++++- .../v2/InputAudioConfigOrBuilder.java | 23 +- .../dialogflow/v2/OutputAudioConfig.java | 4 +- .../dialogflow/v2/OutputAudioEncoding.java | 2 +- .../cloud/dialogflow/v2/SessionProto.java | 62 ++--- .../dialogflow/v2/SpeechModelVariant.java | 214 +++++++++++++++++ .../cloud/dialogflow/v2/SsmlVoiceGender.java | 2 +- .../cloud/dialogflow/v2/audio_config.proto | 125 +++++++++- .../google/cloud/dialogflow/v2/session.proto | 77 ------- .../dialogflow/v2beta1/AudioConfigProto.java | 78 ++++--- .../dialogflow/v2beta1/InputAudioConfig.java | 141 +++++++++++- .../v2beta1/InputAudioConfigOrBuilder.java | 21 ++ .../v2beta1/OutputAudioEncoding.java | 2 +- .../v2beta1/SpeechModelVariant.java | 216 ++++++++++++++++++ .../dialogflow/v2beta1/SsmlVoiceGender.java | 2 +- .../dialogflow/v2beta1/audio_config.proto | 48 +++- .../google-cloud-dialogflow/synth.metadata | 10 +- 19 files changed, 1057 insertions(+), 205 deletions(-) create mode 100644 google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SpeechModelVariant.java create mode 100644 google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/SpeechModelVariant.java diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioConfigProto.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioConfigProto.java index aad2a2d27800..93097710b5d7 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioConfigProto.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioConfigProto.java @@ -12,6 +12,10 @@ public static void registerAllExtensions(com.google.protobuf.ExtensionRegistry r registerAllExtensions((com.google.protobuf.ExtensionRegistryLite) registry); } + static final com.google.protobuf.Descriptors.Descriptor + internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor; + static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable; static final com.google.protobuf.Descriptors.Descriptor internal_static_google_cloud_dialogflow_v2_VoiceSelectionParams_descriptor; static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable @@ -35,31 +39,47 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { java.lang.String[] descriptorData = { "\n-google/cloud/dialogflow/v2/audio_confi" + "g.proto\022\032google.cloud.dialogflow.v2\032\034goo" - + "gle/api/annotations.proto\"f\n\024VoiceSelect" - + "ionParams\022\014\n\004name\030\001 \001(\t\022@\n\013ssml_gender\030\002" - + " \001(\0162+.google.cloud.dialogflow.v2.SsmlVo" - + "iceGender\"\263\001\n\026SynthesizeSpeechConfig\022\025\n\r" - + "speaking_rate\030\001 \001(\001\022\r\n\005pitch\030\002 \001(\001\022\026\n\016vo" - + "lume_gain_db\030\003 \001(\001\022\032\n\022effects_profile_id" - + "\030\005 \003(\t\022?\n\005voice\030\004 \001(\01320.google.cloud.dia" - + "logflow.v2.VoiceSelectionParams\"\315\001\n\021Outp" - + "utAudioConfig\022G\n\016audio_encoding\030\001 \001(\0162/." - + "google.cloud.dialogflow.v2.OutputAudioEn" - + "coding\022\031\n\021sample_rate_hertz\030\002 \001(\005\022T\n\030syn" - + "thesize_speech_config\030\003 \001(\01322.google.clo" - + "ud.dialogflow.v2.SynthesizeSpeechConfig*" - + "\215\001\n\017SsmlVoiceGender\022!\n\035SSML_VOICE_GENDER" - + "_UNSPECIFIED\020\000\022\032\n\026SSML_VOICE_GENDER_MALE" - + "\020\001\022\034\n\030SSML_VOICE_GENDER_FEMALE\020\002\022\035\n\031SSML" - + "_VOICE_GENDER_NEUTRAL\020\003*\244\001\n\023OutputAudioE" - + "ncoding\022%\n!OUTPUT_AUDIO_ENCODING_UNSPECI" - + "FIED\020\000\022#\n\037OUTPUT_AUDIO_ENCODING_LINEAR_1" - + "6\020\001\022\035\n\031OUTPUT_AUDIO_ENCODING_MP3\020\002\022\"\n\036OU" - + "TPUT_AUDIO_ENCODING_OGG_OPUS\020\003B\237\001\n\036com.g" - + "oogle.cloud.dialogflow.v2B\020AudioConfigPr" - + "otoP\001ZDgoogle.golang.org/genproto/google" - + "apis/cloud/dialogflow/v2;dialogflow\370\001\001\242\002" - + "\002DF\252\002\032Google.Cloud.Dialogflow.V2b\006proto3" + + "gle/api/annotations.proto\"\344\001\n\020InputAudio" + + "Config\022A\n\016audio_encoding\030\001 \001(\0162).google." + + "cloud.dialogflow.v2.AudioEncoding\022\031\n\021sam" + + "ple_rate_hertz\030\002 \001(\005\022\025\n\rlanguage_code\030\003 " + + "\001(\t\022\024\n\014phrase_hints\030\004 \003(\t\022E\n\rmodel_varia" + + "nt\030\n \001(\0162..google.cloud.dialogflow.v2.Sp" + + "eechModelVariant\"f\n\024VoiceSelectionParams" + + "\022\014\n\004name\030\001 \001(\t\022@\n\013ssml_gender\030\002 \001(\0162+.go" + + "ogle.cloud.dialogflow.v2.SsmlVoiceGender" + + "\"\263\001\n\026SynthesizeSpeechConfig\022\025\n\rspeaking_" + + "rate\030\001 \001(\001\022\r\n\005pitch\030\002 \001(\001\022\026\n\016volume_gain" + + "_db\030\003 \001(\001\022\032\n\022effects_profile_id\030\005 \003(\t\022?\n" + + "\005voice\030\004 \001(\01320.google.cloud.dialogflow.v" + + "2.VoiceSelectionParams\"\315\001\n\021OutputAudioCo" + + "nfig\022G\n\016audio_encoding\030\001 \001(\0162/.google.cl" + + "oud.dialogflow.v2.OutputAudioEncoding\022\031\n" + + "\021sample_rate_hertz\030\002 \001(\005\022T\n\030synthesize_s" + + "peech_config\030\003 \001(\01322.google.cloud.dialog" + + "flow.v2.SynthesizeSpeechConfig*\373\001\n\rAudio" + + "Encoding\022\036\n\032AUDIO_ENCODING_UNSPECIFIED\020\000" + + "\022\034\n\030AUDIO_ENCODING_LINEAR_16\020\001\022\027\n\023AUDIO_" + + "ENCODING_FLAC\020\002\022\030\n\024AUDIO_ENCODING_MULAW\020" + + "\003\022\026\n\022AUDIO_ENCODING_AMR\020\004\022\031\n\025AUDIO_ENCOD" + + "ING_AMR_WB\020\005\022\033\n\027AUDIO_ENCODING_OGG_OPUS\020" + + "\006\022)\n%AUDIO_ENCODING_SPEEX_WITH_HEADER_BY" + + "TE\020\007*v\n\022SpeechModelVariant\022$\n SPEECH_MOD" + + "EL_VARIANT_UNSPECIFIED\020\000\022\026\n\022USE_BEST_AVA" + + "ILABLE\020\001\022\020\n\014USE_STANDARD\020\002\022\020\n\014USE_ENHANC" + + "ED\020\003*\215\001\n\017SsmlVoiceGender\022!\n\035SSML_VOICE_G" + + "ENDER_UNSPECIFIED\020\000\022\032\n\026SSML_VOICE_GENDER" + + "_MALE\020\001\022\034\n\030SSML_VOICE_GENDER_FEMALE\020\002\022\035\n" + + "\031SSML_VOICE_GENDER_NEUTRAL\020\003*\244\001\n\023OutputA" + + "udioEncoding\022%\n!OUTPUT_AUDIO_ENCODING_UN" + + "SPECIFIED\020\000\022#\n\037OUTPUT_AUDIO_ENCODING_LIN" + + "EAR_16\020\001\022\035\n\031OUTPUT_AUDIO_ENCODING_MP3\020\002\022" + + "\"\n\036OUTPUT_AUDIO_ENCODING_OGG_OPUS\020\003B\237\001\n\036" + + "com.google.cloud.dialogflow.v2B\020AudioCon" + + "figProtoP\001ZDgoogle.golang.org/genproto/g" + + "oogleapis/cloud/dialogflow/v2;dialogflow" + + "\370\001\001\242\002\002DF\252\002\032Google.Cloud.Dialogflow.V2b\006p" + + "roto3" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { @@ -75,8 +95,16 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors( com.google.api.AnnotationsProto.getDescriptor(), }, assigner); - internal_static_google_cloud_dialogflow_v2_VoiceSelectionParams_descriptor = + internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor = getDescriptor().getMessageTypes().get(0); + internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable = + new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor, + new java.lang.String[] { + "AudioEncoding", "SampleRateHertz", "LanguageCode", "PhraseHints", "ModelVariant", + }); + internal_static_google_cloud_dialogflow_v2_VoiceSelectionParams_descriptor = + getDescriptor().getMessageTypes().get(1); internal_static_google_cloud_dialogflow_v2_VoiceSelectionParams_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dialogflow_v2_VoiceSelectionParams_descriptor, @@ -84,7 +112,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors( "Name", "SsmlGender", }); internal_static_google_cloud_dialogflow_v2_SynthesizeSpeechConfig_descriptor = - getDescriptor().getMessageTypes().get(1); + getDescriptor().getMessageTypes().get(2); internal_static_google_cloud_dialogflow_v2_SynthesizeSpeechConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dialogflow_v2_SynthesizeSpeechConfig_descriptor, @@ -92,7 +120,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors( "SpeakingRate", "Pitch", "VolumeGainDb", "EffectsProfileId", "Voice", }); internal_static_google_cloud_dialogflow_v2_OutputAudioConfig_descriptor = - getDescriptor().getMessageTypes().get(2); + getDescriptor().getMessageTypes().get(3); internal_static_google_cloud_dialogflow_v2_OutputAudioConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dialogflow_v2_OutputAudioConfig_descriptor, diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioEncoding.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioEncoding.java index 29d255466510..abf8ded84550 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioEncoding.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioEncoding.java @@ -1,5 +1,5 @@ // Generated by the protocol buffer compiler. DO NOT EDIT! -// source: google/cloud/dialogflow/v2/session.proto +// source: google/cloud/dialogflow/v2/audio_config.proto package com.google.cloud.dialogflow.v2; @@ -274,7 +274,7 @@ public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType } public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { - return com.google.cloud.dialogflow.v2.SessionProto.getDescriptor().getEnumTypes().get(0); + return com.google.cloud.dialogflow.v2.AudioConfigProto.getDescriptor().getEnumTypes().get(0); } private static final AudioEncoding[] VALUES = values(); diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfig.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfig.java index c361c911cb1e..67a572c50889 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfig.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfig.java @@ -1,5 +1,5 @@ // Generated by the protocol buffer compiler. DO NOT EDIT! -// source: google/cloud/dialogflow/v2/session.proto +// source: google/cloud/dialogflow/v2/audio_config.proto package com.google.cloud.dialogflow.v2; @@ -26,6 +26,7 @@ private InputAudioConfig() { audioEncoding_ = 0; languageCode_ = ""; phraseHints_ = com.google.protobuf.LazyStringArrayList.EMPTY; + modelVariant_ = 0; } @java.lang.Override @@ -81,6 +82,13 @@ private InputAudioConfig( phraseHints_.add(s); break; } + case 80: + { + int rawValue = input.readEnum(); + + modelVariant_ = rawValue; + break; + } default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { @@ -104,14 +112,14 @@ private InputAudioConfig( } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return com.google.cloud.dialogflow.v2.SessionProto + return com.google.cloud.dialogflow.v2.AudioConfigProto .internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor; } @java.lang.Override protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { - return com.google.cloud.dialogflow.v2.SessionProto + return com.google.cloud.dialogflow.v2.AudioConfigProto .internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable .ensureFieldAccessorsInitialized( com.google.cloud.dialogflow.v2.InputAudioConfig.class, @@ -290,6 +298,36 @@ public com.google.protobuf.ByteString getPhraseHintsBytes(int index) { return phraseHints_.getByteString(index); } + public static final int MODEL_VARIANT_FIELD_NUMBER = 10; + private int modelVariant_; + /** + * + * + *
+   * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
+   * 
+ * + * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; + */ + public int getModelVariantValue() { + return modelVariant_; + } + /** + * + * + *
+   * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
+   * 
+ * + * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; + */ + public com.google.cloud.dialogflow.v2.SpeechModelVariant getModelVariant() { + @SuppressWarnings("deprecation") + com.google.cloud.dialogflow.v2.SpeechModelVariant result = + com.google.cloud.dialogflow.v2.SpeechModelVariant.valueOf(modelVariant_); + return result == null ? com.google.cloud.dialogflow.v2.SpeechModelVariant.UNRECOGNIZED : result; + } + private byte memoizedIsInitialized = -1; @java.lang.Override @@ -317,6 +355,11 @@ public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io for (int i = 0; i < phraseHints_.size(); i++) { com.google.protobuf.GeneratedMessageV3.writeString(output, 4, phraseHints_.getRaw(i)); } + if (modelVariant_ + != com.google.cloud.dialogflow.v2.SpeechModelVariant.SPEECH_MODEL_VARIANT_UNSPECIFIED + .getNumber()) { + output.writeEnum(10, modelVariant_); + } unknownFields.writeTo(output); } @@ -344,6 +387,11 @@ public int getSerializedSize() { size += dataSize; size += 1 * getPhraseHintsList().size(); } + if (modelVariant_ + != com.google.cloud.dialogflow.v2.SpeechModelVariant.SPEECH_MODEL_VARIANT_UNSPECIFIED + .getNumber()) { + size += com.google.protobuf.CodedOutputStream.computeEnumSize(10, modelVariant_); + } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; @@ -364,6 +412,7 @@ public boolean equals(final java.lang.Object obj) { if (getSampleRateHertz() != other.getSampleRateHertz()) return false; if (!getLanguageCode().equals(other.getLanguageCode())) return false; if (!getPhraseHintsList().equals(other.getPhraseHintsList())) return false; + if (modelVariant_ != other.modelVariant_) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @@ -385,6 +434,8 @@ public int hashCode() { hash = (37 * hash) + PHRASE_HINTS_FIELD_NUMBER; hash = (53 * hash) + getPhraseHintsList().hashCode(); } + hash = (37 * hash) + MODEL_VARIANT_FIELD_NUMBER; + hash = (53 * hash) + modelVariant_; hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; @@ -499,14 +550,14 @@ public static final class Builder extends com.google.protobuf.GeneratedMessageV3 // @@protoc_insertion_point(builder_implements:google.cloud.dialogflow.v2.InputAudioConfig) com.google.cloud.dialogflow.v2.InputAudioConfigOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return com.google.cloud.dialogflow.v2.SessionProto + return com.google.cloud.dialogflow.v2.AudioConfigProto .internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor; } @java.lang.Override protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { - return com.google.cloud.dialogflow.v2.SessionProto + return com.google.cloud.dialogflow.v2.AudioConfigProto .internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable .ensureFieldAccessorsInitialized( com.google.cloud.dialogflow.v2.InputAudioConfig.class, @@ -538,12 +589,14 @@ public Builder clear() { phraseHints_ = com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000008); + modelVariant_ = 0; + return this; } @java.lang.Override public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { - return com.google.cloud.dialogflow.v2.SessionProto + return com.google.cloud.dialogflow.v2.AudioConfigProto .internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor; } @@ -575,6 +628,7 @@ public com.google.cloud.dialogflow.v2.InputAudioConfig buildPartial() { bitField0_ = (bitField0_ & ~0x00000008); } result.phraseHints_ = phraseHints_; + result.modelVariant_ = modelVariant_; result.bitField0_ = to_bitField0_; onBuilt(); return result; @@ -646,6 +700,9 @@ public Builder mergeFrom(com.google.cloud.dialogflow.v2.InputAudioConfig other) } onChanged(); } + if (other.modelVariant_ != 0) { + setModelVariantValue(other.getModelVariantValue()); + } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; @@ -1111,6 +1168,84 @@ public Builder addPhraseHintsBytes(com.google.protobuf.ByteString value) { return this; } + private int modelVariant_ = 0; + /** + * + * + *
+     * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
+     * 
+ * + * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; + */ + public int getModelVariantValue() { + return modelVariant_; + } + /** + * + * + *
+     * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
+     * 
+ * + * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; + */ + public Builder setModelVariantValue(int value) { + modelVariant_ = value; + onChanged(); + return this; + } + /** + * + * + *
+     * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
+     * 
+ * + * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; + */ + public com.google.cloud.dialogflow.v2.SpeechModelVariant getModelVariant() { + @SuppressWarnings("deprecation") + com.google.cloud.dialogflow.v2.SpeechModelVariant result = + com.google.cloud.dialogflow.v2.SpeechModelVariant.valueOf(modelVariant_); + return result == null + ? com.google.cloud.dialogflow.v2.SpeechModelVariant.UNRECOGNIZED + : result; + } + /** + * + * + *
+     * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
+     * 
+ * + * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; + */ + public Builder setModelVariant(com.google.cloud.dialogflow.v2.SpeechModelVariant value) { + if (value == null) { + throw new NullPointerException(); + } + + modelVariant_ = value.getNumber(); + onChanged(); + return this; + } + /** + * + * + *
+     * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
+     * 
+ * + * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; + */ + public Builder clearModelVariant() { + + modelVariant_ = 0; + onChanged(); + return this; + } + @java.lang.Override public final Builder setUnknownFields(final com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfigOrBuilder.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfigOrBuilder.java index f1b0da5bf5fc..c8d0655150a7 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfigOrBuilder.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfigOrBuilder.java @@ -1,5 +1,5 @@ // Generated by the protocol buffer compiler. DO NOT EDIT! -// source: google/cloud/dialogflow/v2/session.proto +// source: google/cloud/dialogflow/v2/audio_config.proto package com.google.cloud.dialogflow.v2; @@ -133,4 +133,25 @@ public interface InputAudioConfigOrBuilder * repeated string phrase_hints = 4; */ com.google.protobuf.ByteString getPhraseHintsBytes(int index); + + /** + * + * + *
+   * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
+   * 
+ * + * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; + */ + int getModelVariantValue(); + /** + * + * + *
+   * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
+   * 
+ * + * .google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10; + */ + com.google.cloud.dialogflow.v2.SpeechModelVariant getModelVariant(); } diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioConfig.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioConfig.java index 3f0a1b4890e2..16b8c21fa7d7 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioConfig.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioConfig.java @@ -7,7 +7,7 @@ * * *
- * Instructs the speech synthesizer how to generate the output audio content.
+ * Instructs the speech synthesizer on how to generate the output audio content.
  * 
* * Protobuf type {@code google.cloud.dialogflow.v2.OutputAudioConfig} @@ -397,7 +397,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * Instructs the speech synthesizer how to generate the output audio content.
+   * Instructs the speech synthesizer on how to generate the output audio content.
    * 
* * Protobuf type {@code google.cloud.dialogflow.v2.OutputAudioConfig} diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioEncoding.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioEncoding.java index c03b8e0ba80e..16c0392a621b 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioEncoding.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioEncoding.java @@ -156,7 +156,7 @@ public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType } public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { - return com.google.cloud.dialogflow.v2.AudioConfigProto.getDescriptor().getEnumTypes().get(1); + return com.google.cloud.dialogflow.v2.AudioConfigProto.getDescriptor().getEnumTypes().get(3); } private static final OutputAudioEncoding[] VALUES = values(); diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SessionProto.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SessionProto.java index b93b98ec2e0b..c579791b0940 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SessionProto.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SessionProto.java @@ -44,10 +44,6 @@ public static void registerAllExtensions(com.google.protobuf.ExtensionRegistry r internal_static_google_cloud_dialogflow_v2_StreamingRecognitionResult_descriptor; static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_google_cloud_dialogflow_v2_StreamingRecognitionResult_fieldAccessorTable; - static final com.google.protobuf.Descriptors.Descriptor - internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor; - static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable; static final com.google.protobuf.Descriptors.Descriptor internal_static_google_cloud_dialogflow_v2_TextInput_descriptor; static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable @@ -153,11 +149,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "ype\022\022\n\ntranscript\030\002 \001(\t\022\020\n\010is_final\030\003 \001(" + "\010\022\022\n\nconfidence\030\004 \001(\002\"X\n\013MessageType\022\034\n\030" + "MESSAGE_TYPE_UNSPECIFIED\020\000\022\016\n\nTRANSCRIPT" - + "\020\001\022\033\n\027END_OF_SINGLE_UTTERANCE\020\002\"\235\001\n\020Inpu" - + "tAudioConfig\022A\n\016audio_encoding\030\001 \001(\0162).g" - + "oogle.cloud.dialogflow.v2.AudioEncoding\022" - + "\031\n\021sample_rate_hertz\030\002 \001(\005\022\025\n\rlanguage_c" - + "ode\030\003 \001(\t\022\024\n\014phrase_hints\030\004 \003(\t\"0\n\tTextI" + + "\020\001\022\033\n\027END_OF_SINGLE_UTTERANCE\020\002\"0\n\tTextI" + "nput\022\014\n\004text\030\001 \001(\t\022\025\n\rlanguage_code\030\002 \001(" + "\t\"^\n\nEventInput\022\014\n\004name\030\001 \001(\t\022+\n\nparamet" + "ers\030\002 \001(\0132\027.google.protobuf.Struct\022\025\n\rla" @@ -166,26 +158,20 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { + "ent\030\001 \001(\010\"^\n\027SentimentAnalysisResult\022C\n\024" + "query_text_sentiment\030\001 \001(\0132%.google.clou" + "d.dialogflow.v2.Sentiment\"-\n\tSentiment\022\r" - + "\n\005score\030\001 \001(\002\022\021\n\tmagnitude\030\002 \001(\002*\373\001\n\rAud" - + "ioEncoding\022\036\n\032AUDIO_ENCODING_UNSPECIFIED" - + "\020\000\022\034\n\030AUDIO_ENCODING_LINEAR_16\020\001\022\027\n\023AUDI" - + "O_ENCODING_FLAC\020\002\022\030\n\024AUDIO_ENCODING_MULA" - + "W\020\003\022\026\n\022AUDIO_ENCODING_AMR\020\004\022\031\n\025AUDIO_ENC" - + "ODING_AMR_WB\020\005\022\033\n\027AUDIO_ENCODING_OGG_OPU" - + "S\020\006\022)\n%AUDIO_ENCODING_SPEEX_WITH_HEADER_" - + "BYTE\020\0072\326\002\n\010Sessions\022\264\001\n\014DetectIntent\022/.g" - + "oogle.cloud.dialogflow.v2.DetectIntentRe" - + "quest\0320.google.cloud.dialogflow.v2.Detec" - + "tIntentResponse\"A\202\323\344\223\002;\"6/v2/{session=pr" - + "ojects/*/agent/sessions/*}:detectIntent:" - + "\001*\022\222\001\n\025StreamingDetectIntent\0228.google.cl" - + "oud.dialogflow.v2.StreamingDetectIntentR" - + "equest\0329.google.cloud.dialogflow.v2.Stre" - + "amingDetectIntentResponse\"\000(\0010\001B\233\001\n\036com." - + "google.cloud.dialogflow.v2B\014SessionProto" - + "P\001ZDgoogle.golang.org/genproto/googleapi" - + "s/cloud/dialogflow/v2;dialogflow\370\001\001\242\002\002DF" - + "\252\002\032Google.Cloud.Dialogflow.V2b\006proto3" + + "\n\005score\030\001 \001(\002\022\021\n\tmagnitude\030\002 \001(\0022\326\002\n\010Ses" + + "sions\022\264\001\n\014DetectIntent\022/.google.cloud.di" + + "alogflow.v2.DetectIntentRequest\0320.google" + + ".cloud.dialogflow.v2.DetectIntentRespons" + + "e\"A\202\323\344\223\002;\"6/v2/{session=projects/*/agent" + + "/sessions/*}:detectIntent:\001*\022\222\001\n\025Streami" + + "ngDetectIntent\0228.google.cloud.dialogflow" + + ".v2.StreamingDetectIntentRequest\0329.googl" + + "e.cloud.dialogflow.v2.StreamingDetectInt" + + "entResponse\"\000(\0010\001B\233\001\n\036com.google.cloud.d" + + "ialogflow.v2B\014SessionProtoP\001ZDgoogle.gol" + + "ang.org/genproto/googleapis/cloud/dialog" + + "flow/v2;dialogflow\370\001\001\242\002\002DF\252\002\032Google.Clou" + + "d.Dialogflow.V2b\006proto3" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { @@ -303,16 +289,8 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors( new java.lang.String[] { "MessageType", "Transcript", "IsFinal", "Confidence", }); - internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor = - getDescriptor().getMessageTypes().get(8); - internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable = - new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( - internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor, - new java.lang.String[] { - "AudioEncoding", "SampleRateHertz", "LanguageCode", "PhraseHints", - }); internal_static_google_cloud_dialogflow_v2_TextInput_descriptor = - getDescriptor().getMessageTypes().get(9); + getDescriptor().getMessageTypes().get(8); internal_static_google_cloud_dialogflow_v2_TextInput_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dialogflow_v2_TextInput_descriptor, @@ -320,7 +298,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors( "Text", "LanguageCode", }); internal_static_google_cloud_dialogflow_v2_EventInput_descriptor = - getDescriptor().getMessageTypes().get(10); + getDescriptor().getMessageTypes().get(9); internal_static_google_cloud_dialogflow_v2_EventInput_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dialogflow_v2_EventInput_descriptor, @@ -328,7 +306,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors( "Name", "Parameters", "LanguageCode", }); internal_static_google_cloud_dialogflow_v2_SentimentAnalysisRequestConfig_descriptor = - getDescriptor().getMessageTypes().get(11); + getDescriptor().getMessageTypes().get(10); internal_static_google_cloud_dialogflow_v2_SentimentAnalysisRequestConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dialogflow_v2_SentimentAnalysisRequestConfig_descriptor, @@ -336,7 +314,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors( "AnalyzeQueryTextSentiment", }); internal_static_google_cloud_dialogflow_v2_SentimentAnalysisResult_descriptor = - getDescriptor().getMessageTypes().get(12); + getDescriptor().getMessageTypes().get(11); internal_static_google_cloud_dialogflow_v2_SentimentAnalysisResult_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dialogflow_v2_SentimentAnalysisResult_descriptor, @@ -344,7 +322,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors( "QueryTextSentiment", }); internal_static_google_cloud_dialogflow_v2_Sentiment_descriptor = - getDescriptor().getMessageTypes().get(13); + getDescriptor().getMessageTypes().get(12); internal_static_google_cloud_dialogflow_v2_Sentiment_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dialogflow_v2_Sentiment_descriptor, diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SpeechModelVariant.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SpeechModelVariant.java new file mode 100644 index 000000000000..ffaf71bf0ecc --- /dev/null +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SpeechModelVariant.java @@ -0,0 +1,214 @@ +// Generated by the protocol buffer compiler. DO NOT EDIT! +// source: google/cloud/dialogflow/v2/audio_config.proto + +package com.google.cloud.dialogflow.v2; + +/** + * + * + *
+ * Variant of the specified [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.
+ * See the [Cloud Speech
+ * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models)
+ * for which models have different variants. For example, the "phone_call" model
+ * has both a standard and an enhanced variant. When you use an enhanced model,
+ * you will generally receive higher quality results than for a standard model.
+ * 
+ * + * Protobuf enum {@code google.cloud.dialogflow.v2.SpeechModelVariant} + */ +public enum SpeechModelVariant implements com.google.protobuf.ProtocolMessageEnum { + /** + * + * + *
+   * No model variant specified. In this case Dialogflow defaults to
+   * USE_BEST_AVAILABLE.
+   * 
+ * + * SPEECH_MODEL_VARIANT_UNSPECIFIED = 0; + */ + SPEECH_MODEL_VARIANT_UNSPECIFIED(0), + /** + * + * + *
+   * Use the best available variant of the [Speech
+   * model][InputAudioConfig.model] that the caller is eligible for.
+   * Please see the [Dialogflow
+   * docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) for
+   * how to make your project eligible for enhanced models.
+   * 
+ * + * USE_BEST_AVAILABLE = 1; + */ + USE_BEST_AVAILABLE(1), + /** + * + * + *
+   * Use standard model variant even if an enhanced model is available.  See the
+   * [Cloud Speech
+   * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models)
+   * for details about enhanced models.
+   * 
+ * + * USE_STANDARD = 2; + */ + USE_STANDARD(2), + /** + * + * + *
+   * Use an enhanced model variant:
+   * * If an enhanced variant does not exist for the given
+   *   [model][google.cloud.dialogflow.v2.InputAudioConfig.model] and request language, Dialogflow falls
+   *   back to the standard variant.
+   *   The [Cloud Speech
+   *   documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models)
+   *   describes which models have enhanced variants.
+   * * If the API caller isn't eligible for enhanced models, Dialogflow returns
+   *   an error. Please see the [Dialogflow
+   *   docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging)
+   *   for how to make your project eligible.
+   * 
+ * + * USE_ENHANCED = 3; + */ + USE_ENHANCED(3), + UNRECOGNIZED(-1), + ; + + /** + * + * + *
+   * No model variant specified. In this case Dialogflow defaults to
+   * USE_BEST_AVAILABLE.
+   * 
+ * + * SPEECH_MODEL_VARIANT_UNSPECIFIED = 0; + */ + public static final int SPEECH_MODEL_VARIANT_UNSPECIFIED_VALUE = 0; + /** + * + * + *
+   * Use the best available variant of the [Speech
+   * model][InputAudioConfig.model] that the caller is eligible for.
+   * Please see the [Dialogflow
+   * docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) for
+   * how to make your project eligible for enhanced models.
+   * 
+ * + * USE_BEST_AVAILABLE = 1; + */ + public static final int USE_BEST_AVAILABLE_VALUE = 1; + /** + * + * + *
+   * Use standard model variant even if an enhanced model is available.  See the
+   * [Cloud Speech
+   * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models)
+   * for details about enhanced models.
+   * 
+ * + * USE_STANDARD = 2; + */ + public static final int USE_STANDARD_VALUE = 2; + /** + * + * + *
+   * Use an enhanced model variant:
+   * * If an enhanced variant does not exist for the given
+   *   [model][google.cloud.dialogflow.v2.InputAudioConfig.model] and request language, Dialogflow falls
+   *   back to the standard variant.
+   *   The [Cloud Speech
+   *   documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models)
+   *   describes which models have enhanced variants.
+   * * If the API caller isn't eligible for enhanced models, Dialogflow returns
+   *   an error. Please see the [Dialogflow
+   *   docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging)
+   *   for how to make your project eligible.
+   * 
+ * + * USE_ENHANCED = 3; + */ + public static final int USE_ENHANCED_VALUE = 3; + + public final int getNumber() { + if (this == UNRECOGNIZED) { + throw new java.lang.IllegalArgumentException( + "Can't get the number of an unknown enum value."); + } + return value; + } + + /** @deprecated Use {@link #forNumber(int)} instead. */ + @java.lang.Deprecated + public static SpeechModelVariant valueOf(int value) { + return forNumber(value); + } + + public static SpeechModelVariant forNumber(int value) { + switch (value) { + case 0: + return SPEECH_MODEL_VARIANT_UNSPECIFIED; + case 1: + return USE_BEST_AVAILABLE; + case 2: + return USE_STANDARD; + case 3: + return USE_ENHANCED; + default: + return null; + } + } + + public static com.google.protobuf.Internal.EnumLiteMap internalGetValueMap() { + return internalValueMap; + } + + private static final com.google.protobuf.Internal.EnumLiteMap + internalValueMap = + new com.google.protobuf.Internal.EnumLiteMap() { + public SpeechModelVariant findValueByNumber(int number) { + return SpeechModelVariant.forNumber(number); + } + }; + + public final com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { + return getDescriptor().getValues().get(ordinal()); + } + + public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { + return getDescriptor(); + } + + public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { + return com.google.cloud.dialogflow.v2.AudioConfigProto.getDescriptor().getEnumTypes().get(1); + } + + private static final SpeechModelVariant[] VALUES = values(); + + public static SpeechModelVariant valueOf( + com.google.protobuf.Descriptors.EnumValueDescriptor desc) { + if (desc.getType() != getDescriptor()) { + throw new java.lang.IllegalArgumentException("EnumValueDescriptor is not for this type."); + } + if (desc.getIndex() == -1) { + return UNRECOGNIZED; + } + return VALUES[desc.getIndex()]; + } + + private final int value; + + private SpeechModelVariant(int value) { + this.value = value; + } + + // @@protoc_insertion_point(enum_scope:google.cloud.dialogflow.v2.SpeechModelVariant) +} diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SsmlVoiceGender.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SsmlVoiceGender.java index 0708caefdbfe..e9108184f68f 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SsmlVoiceGender.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SsmlVoiceGender.java @@ -149,7 +149,7 @@ public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType } public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { - return com.google.cloud.dialogflow.v2.AudioConfigProto.getDescriptor().getEnumTypes().get(0); + return com.google.cloud.dialogflow.v2.AudioConfigProto.getDescriptor().getEnumTypes().get(2); } private static final SsmlVoiceGender[] VALUES = values(); diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/proto/google/cloud/dialogflow/v2/audio_config.proto b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/proto/google/cloud/dialogflow/v2/audio_config.proto index d3ae64e9f8cd..800e5c6ad662 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/proto/google/cloud/dialogflow/v2/audio_config.proto +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/proto/google/cloud/dialogflow/v2/audio_config.proto @@ -27,6 +27,129 @@ option java_outer_classname = "AudioConfigProto"; option java_package = "com.google.cloud.dialogflow.v2"; option objc_class_prefix = "DF"; +// Audio encoding of the audio content sent in the conversational query request. +// Refer to the +// [Cloud Speech API +// documentation](https://cloud.google.com/speech-to-text/docs/basics) for more +// details. +enum AudioEncoding { + // Not specified. + AUDIO_ENCODING_UNSPECIFIED = 0; + + // Uncompressed 16-bit signed little-endian samples (Linear PCM). + AUDIO_ENCODING_LINEAR_16 = 1; + + // [`FLAC`](https://xiph.org/flac/documentation.html) (Free Lossless Audio + // Codec) is the recommended encoding because it is lossless (therefore + // recognition is not compromised) and requires only about half the + // bandwidth of `LINEAR16`. `FLAC` stream encoding supports 16-bit and + // 24-bit samples, however, not all fields in `STREAMINFO` are supported. + AUDIO_ENCODING_FLAC = 2; + + // 8-bit samples that compand 14-bit audio samples using G.711 PCMU/mu-law. + AUDIO_ENCODING_MULAW = 3; + + // Adaptive Multi-Rate Narrowband codec. `sample_rate_hertz` must be 8000. + AUDIO_ENCODING_AMR = 4; + + // Adaptive Multi-Rate Wideband codec. `sample_rate_hertz` must be 16000. + AUDIO_ENCODING_AMR_WB = 5; + + // Opus encoded audio frames in Ogg container + // ([OggOpus](https://wiki.xiph.org/OggOpus)). + // `sample_rate_hertz` must be 16000. + AUDIO_ENCODING_OGG_OPUS = 6; + + // Although the use of lossy encodings is not recommended, if a very low + // bitrate encoding is required, `OGG_OPUS` is highly preferred over + // Speex encoding. The [Speex](https://speex.org/) encoding supported by + // Dialogflow API has a header byte in each block, as in MIME type + // `audio/x-speex-with-header-byte`. + // It is a variant of the RTP Speex encoding defined in + // [RFC 5574](https://tools.ietf.org/html/rfc5574). + // The stream is a sequence of blocks, one block per RTP packet. Each block + // starts with a byte containing the length of the block, in bytes, followed + // by one or more frames of Speex data, padded to an integral number of + // bytes (octets) as specified in RFC 5574. In other words, each RTP header + // is replaced with a single byte containing the block length. Only Speex + // wideband is supported. `sample_rate_hertz` must be 16000. + AUDIO_ENCODING_SPEEX_WITH_HEADER_BYTE = 7; +} + +// Variant of the specified [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use. +// +// See the [Cloud Speech +// documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) +// for which models have different variants. For example, the "phone_call" model +// has both a standard and an enhanced variant. When you use an enhanced model, +// you will generally receive higher quality results than for a standard model. +enum SpeechModelVariant { + // No model variant specified. In this case Dialogflow defaults to + // USE_BEST_AVAILABLE. + SPEECH_MODEL_VARIANT_UNSPECIFIED = 0; + + // Use the best available variant of the [Speech + // model][InputAudioConfig.model] that the caller is eligible for. + // + // Please see the [Dialogflow + // docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) for + // how to make your project eligible for enhanced models. + USE_BEST_AVAILABLE = 1; + + // Use standard model variant even if an enhanced model is available. See the + // [Cloud Speech + // documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + // for details about enhanced models. + USE_STANDARD = 2; + + // Use an enhanced model variant: + // + // * If an enhanced variant does not exist for the given + // [model][google.cloud.dialogflow.v2.InputAudioConfig.model] and request language, Dialogflow falls + // back to the standard variant. + // + // The [Cloud Speech + // documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + // describes which models have enhanced variants. + // + // * If the API caller isn't eligible for enhanced models, Dialogflow returns + // an error. Please see the [Dialogflow + // docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) + // for how to make your project eligible. + USE_ENHANCED = 3; +} + +// Instructs the speech recognizer how to process the audio content. +message InputAudioConfig { + // Required. Audio encoding of the audio content to process. + AudioEncoding audio_encoding = 1; + + // Required. Sample rate (in Hertz) of the audio content sent in the query. + // Refer to + // [Cloud Speech API + // documentation](https://cloud.google.com/speech-to-text/docs/basics) for + // more details. + int32 sample_rate_hertz = 2; + + // Required. The language of the supplied audio. Dialogflow does not do + // translations. See [Language + // Support](https://cloud.google.com/dialogflow-enterprise/docs/reference/language) + // for a list of the currently supported language codes. Note that queries in + // the same session do not necessarily need to specify the same language. + string language_code = 3; + + // Optional. The collection of phrase hints which are used to boost accuracy + // of speech recognition. + // Refer to + // [Cloud Speech API + // documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints) + // for more details. + repeated string phrase_hints = 4; + + // Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use. + SpeechModelVariant model_variant = 10; +} + // Gender of the voice as described in // [SSML voice element](https://www.w3.org/TR/speech-synthesis11/#edef_voice). enum SsmlVoiceGender { @@ -109,7 +232,7 @@ enum OutputAudioEncoding { OUTPUT_AUDIO_ENCODING_OGG_OPUS = 3; } -// Instructs the speech synthesizer how to generate the output audio content. +// Instructs the speech synthesizer on how to generate the output audio content. message OutputAudioConfig { // Required. Audio encoding of the synthesized audio content. OutputAudioEncoding audio_encoding = 1; diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/proto/google/cloud/dialogflow/v2/session.proto b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/proto/google/cloud/dialogflow/v2/session.proto index e10bb6db836d..a21623baf794 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/proto/google/cloud/dialogflow/v2/session.proto +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/proto/google/cloud/dialogflow/v2/session.proto @@ -418,34 +418,6 @@ message StreamingRecognitionResult { float confidence = 4; } -// Instructs the speech recognizer how to process the audio content. -message InputAudioConfig { - // Required. Audio encoding of the audio content to process. - AudioEncoding audio_encoding = 1; - - // Required. Sample rate (in Hertz) of the audio content sent in the query. - // Refer to - // [Cloud Speech API - // documentation](https://cloud.google.com/speech-to-text/docs/basics) for - // more details. - int32 sample_rate_hertz = 2; - - // Required. The language of the supplied audio. Dialogflow does not do - // translations. See [Language - // Support](https://cloud.google.com/dialogflow-enterprise/docs/reference/language) - // for a list of the currently supported language codes. Note that queries in - // the same session do not necessarily need to specify the same language. - string language_code = 3; - - // Optional. The collection of phrase hints which are used to boost accuracy - // of speech recognition. - // Refer to - // [Cloud Speech API - // documentation](https://cloud.google.com/speech-to-text/docs/basics#phrase-hints) - // for more details. - repeated string phrase_hints = 4; -} - // Represents the natural language text to be processed. message TextInput { // Required. The UTF-8 encoded natural language text to be processed. @@ -459,55 +431,6 @@ message TextInput { string language_code = 2; } -// Audio encoding of the audio content sent in the conversational query request. -// Refer to the -// [Cloud Speech API -// documentation](https://cloud.google.com/speech-to-text/docs/basics) for more -// details. -enum AudioEncoding { - // Not specified. - AUDIO_ENCODING_UNSPECIFIED = 0; - - // Uncompressed 16-bit signed little-endian samples (Linear PCM). - AUDIO_ENCODING_LINEAR_16 = 1; - - // [`FLAC`](https://xiph.org/flac/documentation.html) (Free Lossless Audio - // Codec) is the recommended encoding because it is lossless (therefore - // recognition is not compromised) and requires only about half the - // bandwidth of `LINEAR16`. `FLAC` stream encoding supports 16-bit and - // 24-bit samples, however, not all fields in `STREAMINFO` are supported. - AUDIO_ENCODING_FLAC = 2; - - // 8-bit samples that compand 14-bit audio samples using G.711 PCMU/mu-law. - AUDIO_ENCODING_MULAW = 3; - - // Adaptive Multi-Rate Narrowband codec. `sample_rate_hertz` must be 8000. - AUDIO_ENCODING_AMR = 4; - - // Adaptive Multi-Rate Wideband codec. `sample_rate_hertz` must be 16000. - AUDIO_ENCODING_AMR_WB = 5; - - // Opus encoded audio frames in Ogg container - // ([OggOpus](https://wiki.xiph.org/OggOpus)). - // `sample_rate_hertz` must be 16000. - AUDIO_ENCODING_OGG_OPUS = 6; - - // Although the use of lossy encodings is not recommended, if a very low - // bitrate encoding is required, `OGG_OPUS` is highly preferred over - // Speex encoding. The [Speex](https://speex.org/) encoding supported by - // Dialogflow API has a header byte in each block, as in MIME type - // `audio/x-speex-with-header-byte`. - // It is a variant of the RTP Speex encoding defined in - // [RFC 5574](https://tools.ietf.org/html/rfc5574). - // The stream is a sequence of blocks, one block per RTP packet. Each block - // starts with a byte containing the length of the block, in bytes, followed - // by one or more frames of Speex data, padded to an integral number of - // bytes (octets) as specified in RFC 5574. In other words, each RTP header - // is replaced with a single byte containing the block length. Only Speex - // wideband is supported. `sample_rate_hertz` must be 16000. - AUDIO_ENCODING_SPEEX_WITH_HEADER_BYTE = 7; -} - // Events allow for matching intents by event name instead of the natural // language input. For instance, input `` can trigger a personalized welcome response. diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/AudioConfigProto.java b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/AudioConfigProto.java index 4ef97493f92d..7e67a1d09cf9 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/AudioConfigProto.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/AudioConfigProto.java @@ -39,43 +39,48 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { java.lang.String[] descriptorData = { "\n2google/cloud/dialogflow/v2beta1/audio_" + "config.proto\022\037google.cloud.dialogflow.v2" - + "beta1\032\034google/api/annotations.proto\"\261\001\n\020" + + "beta1\032\034google/api/annotations.proto\"\375\001\n\020" + "InputAudioConfig\022F\n\016audio_encoding\030\001 \001(\016" + "2..google.cloud.dialogflow.v2beta1.Audio" + "Encoding\022\031\n\021sample_rate_hertz\030\002 \001(\005\022\025\n\rl" + "anguage_code\030\003 \001(\t\022\024\n\014phrase_hints\030\004 \003(\t" - + "\022\r\n\005model\030\007 \001(\t\"k\n\024VoiceSelectionParams\022" - + "\014\n\004name\030\001 \001(\t\022E\n\013ssml_gender\030\002 \001(\01620.goo" - + "gle.cloud.dialogflow.v2beta1.SsmlVoiceGe" - + "nder\"\270\001\n\026SynthesizeSpeechConfig\022\025\n\rspeak" - + "ing_rate\030\001 \001(\001\022\r\n\005pitch\030\002 \001(\001\022\026\n\016volume_" - + "gain_db\030\003 \001(\001\022\032\n\022effects_profile_id\030\005 \003(" - + "\t\022D\n\005voice\030\004 \001(\01325.google.cloud.dialogfl" - + "ow.v2beta1.VoiceSelectionParams\"\327\001\n\021Outp" - + "utAudioConfig\022L\n\016audio_encoding\030\001 \001(\01624." - + "google.cloud.dialogflow.v2beta1.OutputAu" - + "dioEncoding\022\031\n\021sample_rate_hertz\030\002 \001(\005\022Y" - + "\n\030synthesize_speech_config\030\003 \001(\01327.googl" - + "e.cloud.dialogflow.v2beta1.SynthesizeSpe" - + "echConfig*\373\001\n\rAudioEncoding\022\036\n\032AUDIO_ENC" - + "ODING_UNSPECIFIED\020\000\022\034\n\030AUDIO_ENCODING_LI" - + "NEAR_16\020\001\022\027\n\023AUDIO_ENCODING_FLAC\020\002\022\030\n\024AU" - + "DIO_ENCODING_MULAW\020\003\022\026\n\022AUDIO_ENCODING_A" - + "MR\020\004\022\031\n\025AUDIO_ENCODING_AMR_WB\020\005\022\033\n\027AUDIO" - + "_ENCODING_OGG_OPUS\020\006\022)\n%AUDIO_ENCODING_S" - + "PEEX_WITH_HEADER_BYTE\020\007*\215\001\n\017SsmlVoiceGen" - + "der\022!\n\035SSML_VOICE_GENDER_UNSPECIFIED\020\000\022\032" - + "\n\026SSML_VOICE_GENDER_MALE\020\001\022\034\n\030SSML_VOICE" - + "_GENDER_FEMALE\020\002\022\035\n\031SSML_VOICE_GENDER_NE" - + "UTRAL\020\003*\244\001\n\023OutputAudioEncoding\022%\n!OUTPU" - + "T_AUDIO_ENCODING_UNSPECIFIED\020\000\022#\n\037OUTPUT" - + "_AUDIO_ENCODING_LINEAR_16\020\001\022\035\n\031OUTPUT_AU" - + "DIO_ENCODING_MP3\020\002\022\"\n\036OUTPUT_AUDIO_ENCOD" - + "ING_OGG_OPUS\020\003B\256\001\n#com.google.cloud.dial" - + "ogflow.v2beta1B\020AudioConfigProtoP\001ZIgoog" - + "le.golang.org/genproto/googleapis/cloud/" - + "dialogflow/v2beta1;dialogflow\370\001\001\242\002\002DF\252\002\037" - + "Google.Cloud.Dialogflow.V2beta1b\006proto3" + + "\022\r\n\005model\030\007 \001(\t\022J\n\rmodel_variant\030\n \001(\01623" + + ".google.cloud.dialogflow.v2beta1.SpeechM" + + "odelVariant\"k\n\024VoiceSelectionParams\022\014\n\004n" + + "ame\030\001 \001(\t\022E\n\013ssml_gender\030\002 \001(\01620.google." + + "cloud.dialogflow.v2beta1.SsmlVoiceGender" + + "\"\270\001\n\026SynthesizeSpeechConfig\022\025\n\rspeaking_" + + "rate\030\001 \001(\001\022\r\n\005pitch\030\002 \001(\001\022\026\n\016volume_gain" + + "_db\030\003 \001(\001\022\032\n\022effects_profile_id\030\005 \003(\t\022D\n" + + "\005voice\030\004 \001(\01325.google.cloud.dialogflow.v" + + "2beta1.VoiceSelectionParams\"\327\001\n\021OutputAu" + + "dioConfig\022L\n\016audio_encoding\030\001 \001(\01624.goog" + + "le.cloud.dialogflow.v2beta1.OutputAudioE" + + "ncoding\022\031\n\021sample_rate_hertz\030\002 \001(\005\022Y\n\030sy" + + "nthesize_speech_config\030\003 \001(\01327.google.cl" + + "oud.dialogflow.v2beta1.SynthesizeSpeechC" + + "onfig*\373\001\n\rAudioEncoding\022\036\n\032AUDIO_ENCODIN" + + "G_UNSPECIFIED\020\000\022\034\n\030AUDIO_ENCODING_LINEAR" + + "_16\020\001\022\027\n\023AUDIO_ENCODING_FLAC\020\002\022\030\n\024AUDIO_" + + "ENCODING_MULAW\020\003\022\026\n\022AUDIO_ENCODING_AMR\020\004" + + "\022\031\n\025AUDIO_ENCODING_AMR_WB\020\005\022\033\n\027AUDIO_ENC" + + "ODING_OGG_OPUS\020\006\022)\n%AUDIO_ENCODING_SPEEX" + + "_WITH_HEADER_BYTE\020\007*v\n\022SpeechModelVarian" + + "t\022$\n SPEECH_MODEL_VARIANT_UNSPECIFIED\020\000\022" + + "\026\n\022USE_BEST_AVAILABLE\020\001\022\020\n\014USE_STANDARD\020" + + "\002\022\020\n\014USE_ENHANCED\020\003*\215\001\n\017SsmlVoiceGender\022" + + "!\n\035SSML_VOICE_GENDER_UNSPECIFIED\020\000\022\032\n\026SS" + + "ML_VOICE_GENDER_MALE\020\001\022\034\n\030SSML_VOICE_GEN" + + "DER_FEMALE\020\002\022\035\n\031SSML_VOICE_GENDER_NEUTRA" + + "L\020\003*\244\001\n\023OutputAudioEncoding\022%\n!OUTPUT_AU" + + "DIO_ENCODING_UNSPECIFIED\020\000\022#\n\037OUTPUT_AUD" + + "IO_ENCODING_LINEAR_16\020\001\022\035\n\031OUTPUT_AUDIO_" + + "ENCODING_MP3\020\002\022\"\n\036OUTPUT_AUDIO_ENCODING_" + + "OGG_OPUS\020\003B\256\001\n#com.google.cloud.dialogfl" + + "ow.v2beta1B\020AudioConfigProtoP\001ZIgoogle.g" + + "olang.org/genproto/googleapis/cloud/dial" + + "ogflow/v2beta1;dialogflow\370\001\001\242\002\002DF\252\002\037Goog" + + "le.Cloud.Dialogflow.V2beta1b\006proto3" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { @@ -97,7 +102,12 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors( new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dialogflow_v2beta1_InputAudioConfig_descriptor, new java.lang.String[] { - "AudioEncoding", "SampleRateHertz", "LanguageCode", "PhraseHints", "Model", + "AudioEncoding", + "SampleRateHertz", + "LanguageCode", + "PhraseHints", + "Model", + "ModelVariant", }); internal_static_google_cloud_dialogflow_v2beta1_VoiceSelectionParams_descriptor = getDescriptor().getMessageTypes().get(1); diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/InputAudioConfig.java b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/InputAudioConfig.java index 55a9a07f01f8..40e1dc84af93 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/InputAudioConfig.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/InputAudioConfig.java @@ -7,7 +7,7 @@ * * *
- * Instructs the speech recognizer how to process the audio content.
+ * Instructs the speech recognizer on how to process the audio content.
  * 
* * Protobuf type {@code google.cloud.dialogflow.v2beta1.InputAudioConfig} @@ -27,6 +27,7 @@ private InputAudioConfig() { languageCode_ = ""; phraseHints_ = com.google.protobuf.LazyStringArrayList.EMPTY; model_ = ""; + modelVariant_ = 0; } @java.lang.Override @@ -89,6 +90,13 @@ private InputAudioConfig( model_ = s; break; } + case 80: + { + int rawValue = input.readEnum(); + + modelVariant_ = rawValue; + break; + } default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { @@ -361,6 +369,38 @@ public com.google.protobuf.ByteString getModelBytes() { } } + public static final int MODEL_VARIANT_FIELD_NUMBER = 10; + private int modelVariant_; + /** + * + * + *
+   * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use.
+   * 
+ * + * .google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10; + */ + public int getModelVariantValue() { + return modelVariant_; + } + /** + * + * + *
+   * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use.
+   * 
+ * + * .google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10; + */ + public com.google.cloud.dialogflow.v2beta1.SpeechModelVariant getModelVariant() { + @SuppressWarnings("deprecation") + com.google.cloud.dialogflow.v2beta1.SpeechModelVariant result = + com.google.cloud.dialogflow.v2beta1.SpeechModelVariant.valueOf(modelVariant_); + return result == null + ? com.google.cloud.dialogflow.v2beta1.SpeechModelVariant.UNRECOGNIZED + : result; + } + private byte memoizedIsInitialized = -1; @java.lang.Override @@ -392,6 +432,11 @@ public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io if (!getModelBytes().isEmpty()) { com.google.protobuf.GeneratedMessageV3.writeString(output, 7, model_); } + if (modelVariant_ + != com.google.cloud.dialogflow.v2beta1.SpeechModelVariant.SPEECH_MODEL_VARIANT_UNSPECIFIED + .getNumber()) { + output.writeEnum(10, modelVariant_); + } unknownFields.writeTo(output); } @@ -423,6 +468,11 @@ public int getSerializedSize() { if (!getModelBytes().isEmpty()) { size += com.google.protobuf.GeneratedMessageV3.computeStringSize(7, model_); } + if (modelVariant_ + != com.google.cloud.dialogflow.v2beta1.SpeechModelVariant.SPEECH_MODEL_VARIANT_UNSPECIFIED + .getNumber()) { + size += com.google.protobuf.CodedOutputStream.computeEnumSize(10, modelVariant_); + } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; @@ -444,6 +494,7 @@ public boolean equals(final java.lang.Object obj) { if (!getLanguageCode().equals(other.getLanguageCode())) return false; if (!getPhraseHintsList().equals(other.getPhraseHintsList())) return false; if (!getModel().equals(other.getModel())) return false; + if (modelVariant_ != other.modelVariant_) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @@ -467,6 +518,8 @@ public int hashCode() { } hash = (37 * hash) + MODEL_FIELD_NUMBER; hash = (53 * hash) + getModel().hashCode(); + hash = (37 * hash) + MODEL_VARIANT_FIELD_NUMBER; + hash = (53 * hash) + modelVariant_; hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; @@ -571,7 +624,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
-   * Instructs the speech recognizer how to process the audio content.
+   * Instructs the speech recognizer on how to process the audio content.
    * 
* * Protobuf type {@code google.cloud.dialogflow.v2beta1.InputAudioConfig} @@ -622,6 +675,8 @@ public Builder clear() { bitField0_ = (bitField0_ & ~0x00000008); model_ = ""; + modelVariant_ = 0; + return this; } @@ -660,6 +715,7 @@ public com.google.cloud.dialogflow.v2beta1.InputAudioConfig buildPartial() { } result.phraseHints_ = phraseHints_; result.model_ = model_; + result.modelVariant_ = modelVariant_; result.bitField0_ = to_bitField0_; onBuilt(); return result; @@ -735,6 +791,9 @@ public Builder mergeFrom(com.google.cloud.dialogflow.v2beta1.InputAudioConfig ot model_ = other.model_; onChanged(); } + if (other.modelVariant_ != 0) { + setModelVariantValue(other.getModelVariantValue()); + } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; @@ -1347,6 +1406,84 @@ public Builder setModelBytes(com.google.protobuf.ByteString value) { return this; } + private int modelVariant_ = 0; + /** + * + * + *
+     * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use.
+     * 
+ * + * .google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10; + */ + public int getModelVariantValue() { + return modelVariant_; + } + /** + * + * + *
+     * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use.
+     * 
+ * + * .google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10; + */ + public Builder setModelVariantValue(int value) { + modelVariant_ = value; + onChanged(); + return this; + } + /** + * + * + *
+     * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use.
+     * 
+ * + * .google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10; + */ + public com.google.cloud.dialogflow.v2beta1.SpeechModelVariant getModelVariant() { + @SuppressWarnings("deprecation") + com.google.cloud.dialogflow.v2beta1.SpeechModelVariant result = + com.google.cloud.dialogflow.v2beta1.SpeechModelVariant.valueOf(modelVariant_); + return result == null + ? com.google.cloud.dialogflow.v2beta1.SpeechModelVariant.UNRECOGNIZED + : result; + } + /** + * + * + *
+     * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use.
+     * 
+ * + * .google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10; + */ + public Builder setModelVariant(com.google.cloud.dialogflow.v2beta1.SpeechModelVariant value) { + if (value == null) { + throw new NullPointerException(); + } + + modelVariant_ = value.getNumber(); + onChanged(); + return this; + } + /** + * + * + *
+     * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use.
+     * 
+ * + * .google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10; + */ + public Builder clearModelVariant() { + + modelVariant_ = 0; + onChanged(); + return this; + } + @java.lang.Override public final Builder setUnknownFields(final com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/InputAudioConfigOrBuilder.java b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/InputAudioConfigOrBuilder.java index a2a2c48b8644..da25c3ef3063 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/InputAudioConfigOrBuilder.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/InputAudioConfigOrBuilder.java @@ -174,4 +174,25 @@ public interface InputAudioConfigOrBuilder * string model = 7; */ com.google.protobuf.ByteString getModelBytes(); + + /** + * + * + *
+   * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use.
+   * 
+ * + * .google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10; + */ + int getModelVariantValue(); + /** + * + * + *
+   * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use.
+   * 
+ * + * .google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10; + */ + com.google.cloud.dialogflow.v2beta1.SpeechModelVariant getModelVariant(); } diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/OutputAudioEncoding.java b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/OutputAudioEncoding.java index f46dee66703a..1de522c1c579 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/OutputAudioEncoding.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/OutputAudioEncoding.java @@ -158,7 +158,7 @@ public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { return com.google.cloud.dialogflow.v2beta1.AudioConfigProto.getDescriptor() .getEnumTypes() - .get(2); + .get(3); } private static final OutputAudioEncoding[] VALUES = values(); diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/SpeechModelVariant.java b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/SpeechModelVariant.java new file mode 100644 index 000000000000..c08c12c25e7c --- /dev/null +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/SpeechModelVariant.java @@ -0,0 +1,216 @@ +// Generated by the protocol buffer compiler. DO NOT EDIT! +// source: google/cloud/dialogflow/v2beta1/audio_config.proto + +package com.google.cloud.dialogflow.v2beta1; + +/** + * + * + *
+ * Variant of the specified [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use.
+ * See the [Cloud Speech
+ * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models)
+ * for which models have different variants. For example, the "phone_call" model
+ * has both a standard and an enhanced variant. When you use an enhanced model,
+ * you will generally receive higher quality results than for a standard model.
+ * 
+ * + * Protobuf enum {@code google.cloud.dialogflow.v2beta1.SpeechModelVariant} + */ +public enum SpeechModelVariant implements com.google.protobuf.ProtocolMessageEnum { + /** + * + * + *
+   * No model variant specified. In this case Dialogflow defaults to
+   * USE_BEST_AVAILABLE.
+   * 
+ * + * SPEECH_MODEL_VARIANT_UNSPECIFIED = 0; + */ + SPEECH_MODEL_VARIANT_UNSPECIFIED(0), + /** + * + * + *
+   * Use the best available variant of the [Speech
+   * model][InputAudioConfig.model] that the caller is eligible for.
+   * Please see the [Dialogflow
+   * docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) for
+   * how to make your project eligible for enhanced models.
+   * 
+ * + * USE_BEST_AVAILABLE = 1; + */ + USE_BEST_AVAILABLE(1), + /** + * + * + *
+   * Use standard model variant even if an enhanced model is available.  See the
+   * [Cloud Speech
+   * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models)
+   * for details about enhanced models.
+   * 
+ * + * USE_STANDARD = 2; + */ + USE_STANDARD(2), + /** + * + * + *
+   * Use an enhanced model variant:
+   * * If an enhanced variant does not exist for the given
+   *   [model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] and request language, Dialogflow falls
+   *   back to the standard variant.
+   *   The [Cloud Speech
+   *   documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models)
+   *   describes which models have enhanced variants.
+   * * If the API caller isn't eligible for enhanced models, Dialogflow returns
+   *   an error.  Please see the [Dialogflow
+   *   docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging)
+   *   for how to make your project eligible.
+   * 
+ * + * USE_ENHANCED = 3; + */ + USE_ENHANCED(3), + UNRECOGNIZED(-1), + ; + + /** + * + * + *
+   * No model variant specified. In this case Dialogflow defaults to
+   * USE_BEST_AVAILABLE.
+   * 
+ * + * SPEECH_MODEL_VARIANT_UNSPECIFIED = 0; + */ + public static final int SPEECH_MODEL_VARIANT_UNSPECIFIED_VALUE = 0; + /** + * + * + *
+   * Use the best available variant of the [Speech
+   * model][InputAudioConfig.model] that the caller is eligible for.
+   * Please see the [Dialogflow
+   * docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) for
+   * how to make your project eligible for enhanced models.
+   * 
+ * + * USE_BEST_AVAILABLE = 1; + */ + public static final int USE_BEST_AVAILABLE_VALUE = 1; + /** + * + * + *
+   * Use standard model variant even if an enhanced model is available.  See the
+   * [Cloud Speech
+   * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models)
+   * for details about enhanced models.
+   * 
+ * + * USE_STANDARD = 2; + */ + public static final int USE_STANDARD_VALUE = 2; + /** + * + * + *
+   * Use an enhanced model variant:
+   * * If an enhanced variant does not exist for the given
+   *   [model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] and request language, Dialogflow falls
+   *   back to the standard variant.
+   *   The [Cloud Speech
+   *   documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models)
+   *   describes which models have enhanced variants.
+   * * If the API caller isn't eligible for enhanced models, Dialogflow returns
+   *   an error.  Please see the [Dialogflow
+   *   docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging)
+   *   for how to make your project eligible.
+   * 
+ * + * USE_ENHANCED = 3; + */ + public static final int USE_ENHANCED_VALUE = 3; + + public final int getNumber() { + if (this == UNRECOGNIZED) { + throw new java.lang.IllegalArgumentException( + "Can't get the number of an unknown enum value."); + } + return value; + } + + /** @deprecated Use {@link #forNumber(int)} instead. */ + @java.lang.Deprecated + public static SpeechModelVariant valueOf(int value) { + return forNumber(value); + } + + public static SpeechModelVariant forNumber(int value) { + switch (value) { + case 0: + return SPEECH_MODEL_VARIANT_UNSPECIFIED; + case 1: + return USE_BEST_AVAILABLE; + case 2: + return USE_STANDARD; + case 3: + return USE_ENHANCED; + default: + return null; + } + } + + public static com.google.protobuf.Internal.EnumLiteMap internalGetValueMap() { + return internalValueMap; + } + + private static final com.google.protobuf.Internal.EnumLiteMap + internalValueMap = + new com.google.protobuf.Internal.EnumLiteMap() { + public SpeechModelVariant findValueByNumber(int number) { + return SpeechModelVariant.forNumber(number); + } + }; + + public final com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { + return getDescriptor().getValues().get(ordinal()); + } + + public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { + return getDescriptor(); + } + + public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { + return com.google.cloud.dialogflow.v2beta1.AudioConfigProto.getDescriptor() + .getEnumTypes() + .get(1); + } + + private static final SpeechModelVariant[] VALUES = values(); + + public static SpeechModelVariant valueOf( + com.google.protobuf.Descriptors.EnumValueDescriptor desc) { + if (desc.getType() != getDescriptor()) { + throw new java.lang.IllegalArgumentException("EnumValueDescriptor is not for this type."); + } + if (desc.getIndex() == -1) { + return UNRECOGNIZED; + } + return VALUES[desc.getIndex()]; + } + + private final int value; + + private SpeechModelVariant(int value) { + this.value = value; + } + + // @@protoc_insertion_point(enum_scope:google.cloud.dialogflow.v2beta1.SpeechModelVariant) +} diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/SsmlVoiceGender.java b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/SsmlVoiceGender.java index 8e2fdfa13796..3e4335d48ae2 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/SsmlVoiceGender.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/SsmlVoiceGender.java @@ -151,7 +151,7 @@ public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { return com.google.cloud.dialogflow.v2beta1.AudioConfigProto.getDescriptor() .getEnumTypes() - .get(1); + .get(2); } private static final SsmlVoiceGender[] VALUES = values(); diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/proto/google/cloud/dialogflow/v2beta1/audio_config.proto b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/proto/google/cloud/dialogflow/v2beta1/audio_config.proto index 8045290b22d4..9ce81863dc05 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/proto/google/cloud/dialogflow/v2beta1/audio_config.proto +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/proto/google/cloud/dialogflow/v2beta1/audio_config.proto @@ -76,7 +76,50 @@ enum AudioEncoding { AUDIO_ENCODING_SPEEX_WITH_HEADER_BYTE = 7; } -// Instructs the speech recognizer how to process the audio content. +// Variant of the specified [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use. +// +// See the [Cloud Speech +// documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) +// for which models have different variants. For example, the "phone_call" model +// has both a standard and an enhanced variant. When you use an enhanced model, +// you will generally receive higher quality results than for a standard model. +enum SpeechModelVariant { + // No model variant specified. In this case Dialogflow defaults to + // USE_BEST_AVAILABLE. + SPEECH_MODEL_VARIANT_UNSPECIFIED = 0; + + // Use the best available variant of the [Speech + // model][InputAudioConfig.model] that the caller is eligible for. + // + // Please see the [Dialogflow + // docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) for + // how to make your project eligible for enhanced models. + USE_BEST_AVAILABLE = 1; + + // Use standard model variant even if an enhanced model is available. See the + // [Cloud Speech + // documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + // for details about enhanced models. + USE_STANDARD = 2; + + // Use an enhanced model variant: + // + // * If an enhanced variant does not exist for the given + // [model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] and request language, Dialogflow falls + // back to the standard variant. + // + // The [Cloud Speech + // documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + // describes which models have enhanced variants. + // + // * If the API caller isn't eligible for enhanced models, Dialogflow returns + // an error. Please see the [Dialogflow + // docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) + // for how to make your project eligible. + USE_ENHANCED = 3; +} + +// Instructs the speech recognizer on how to process the audio content. message InputAudioConfig { // Required. Audio encoding of the audio content to process. AudioEncoding audio_encoding = 1; @@ -115,6 +158,9 @@ message InputAudioConfig { // documentation](https://cloud.google.com/speech-to-text/docs/basics#select-model) // for more details. string model = 7; + + // Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use. + SpeechModelVariant model_variant = 10; } // Gender of the voice as described in diff --git a/google-cloud-clients/google-cloud-dialogflow/synth.metadata b/google-cloud-clients/google-cloud-dialogflow/synth.metadata index c12021cfdfc0..9f0ddf72cbec 100644 --- a/google-cloud-clients/google-cloud-dialogflow/synth.metadata +++ b/google-cloud-clients/google-cloud-dialogflow/synth.metadata @@ -1,19 +1,19 @@ { - "updateTime": "2019-04-10T07:43:29.853673Z", + "updateTime": "2019-04-30T07:42:19.370029Z", "sources": [ { "generator": { "name": "artman", - "version": "0.16.25", - "dockerImage": "googleapis/artman@sha256:d9597f983d1d4e61272c63cb97b7d8f8234da9999526c35d357de3d781f0ec1b" + "version": "0.17.1", + "dockerImage": "googleapis/artman@sha256:a40ca4dd4ef031c0ded4df4909ffdf7b3f20d29b23e682ef991eb60ba0ca6025" } }, { "git": { "name": "googleapis", "remote": "https://github.com/googleapis/googleapis.git", - "sha": "3f57de2fbe951640b587d9c2d503daf554747749", - "internalRef": "242707837" + "sha": "be3005c8a16e8424d0330e977ac78f2976bf01d3", + "internalRef": "245840921" } } ],