diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioConfigProto.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioConfigProto.java index aad2a2d27800..93097710b5d7 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioConfigProto.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioConfigProto.java @@ -12,6 +12,10 @@ public static void registerAllExtensions(com.google.protobuf.ExtensionRegistry r registerAllExtensions((com.google.protobuf.ExtensionRegistryLite) registry); } + static final com.google.protobuf.Descriptors.Descriptor + internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor; + static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable + internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable; static final com.google.protobuf.Descriptors.Descriptor internal_static_google_cloud_dialogflow_v2_VoiceSelectionParams_descriptor; static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable @@ -35,31 +39,47 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { java.lang.String[] descriptorData = { "\n-google/cloud/dialogflow/v2/audio_confi" + "g.proto\022\032google.cloud.dialogflow.v2\032\034goo" - + "gle/api/annotations.proto\"f\n\024VoiceSelect" - + "ionParams\022\014\n\004name\030\001 \001(\t\022@\n\013ssml_gender\030\002" - + " \001(\0162+.google.cloud.dialogflow.v2.SsmlVo" - + "iceGender\"\263\001\n\026SynthesizeSpeechConfig\022\025\n\r" - + "speaking_rate\030\001 \001(\001\022\r\n\005pitch\030\002 \001(\001\022\026\n\016vo" - + "lume_gain_db\030\003 \001(\001\022\032\n\022effects_profile_id" - + "\030\005 \003(\t\022?\n\005voice\030\004 \001(\01320.google.cloud.dia" - + "logflow.v2.VoiceSelectionParams\"\315\001\n\021Outp" - + "utAudioConfig\022G\n\016audio_encoding\030\001 \001(\0162/." - + "google.cloud.dialogflow.v2.OutputAudioEn" - + "coding\022\031\n\021sample_rate_hertz\030\002 \001(\005\022T\n\030syn" - + "thesize_speech_config\030\003 \001(\01322.google.clo" - + "ud.dialogflow.v2.SynthesizeSpeechConfig*" - + "\215\001\n\017SsmlVoiceGender\022!\n\035SSML_VOICE_GENDER" - + "_UNSPECIFIED\020\000\022\032\n\026SSML_VOICE_GENDER_MALE" - + "\020\001\022\034\n\030SSML_VOICE_GENDER_FEMALE\020\002\022\035\n\031SSML" - + "_VOICE_GENDER_NEUTRAL\020\003*\244\001\n\023OutputAudioE" - + "ncoding\022%\n!OUTPUT_AUDIO_ENCODING_UNSPECI" - + "FIED\020\000\022#\n\037OUTPUT_AUDIO_ENCODING_LINEAR_1" - + "6\020\001\022\035\n\031OUTPUT_AUDIO_ENCODING_MP3\020\002\022\"\n\036OU" - + "TPUT_AUDIO_ENCODING_OGG_OPUS\020\003B\237\001\n\036com.g" - + "oogle.cloud.dialogflow.v2B\020AudioConfigPr" - + "otoP\001ZDgoogle.golang.org/genproto/google" - + "apis/cloud/dialogflow/v2;dialogflow\370\001\001\242\002" - + "\002DF\252\002\032Google.Cloud.Dialogflow.V2b\006proto3" + + "gle/api/annotations.proto\"\344\001\n\020InputAudio" + + "Config\022A\n\016audio_encoding\030\001 \001(\0162).google." + + "cloud.dialogflow.v2.AudioEncoding\022\031\n\021sam" + + "ple_rate_hertz\030\002 \001(\005\022\025\n\rlanguage_code\030\003 " + + "\001(\t\022\024\n\014phrase_hints\030\004 \003(\t\022E\n\rmodel_varia" + + "nt\030\n \001(\0162..google.cloud.dialogflow.v2.Sp" + + "eechModelVariant\"f\n\024VoiceSelectionParams" + + "\022\014\n\004name\030\001 \001(\t\022@\n\013ssml_gender\030\002 \001(\0162+.go" + + "ogle.cloud.dialogflow.v2.SsmlVoiceGender" + + "\"\263\001\n\026SynthesizeSpeechConfig\022\025\n\rspeaking_" + + "rate\030\001 \001(\001\022\r\n\005pitch\030\002 \001(\001\022\026\n\016volume_gain" + + "_db\030\003 \001(\001\022\032\n\022effects_profile_id\030\005 \003(\t\022?\n" + + "\005voice\030\004 \001(\01320.google.cloud.dialogflow.v" + + "2.VoiceSelectionParams\"\315\001\n\021OutputAudioCo" + + "nfig\022G\n\016audio_encoding\030\001 \001(\0162/.google.cl" + + "oud.dialogflow.v2.OutputAudioEncoding\022\031\n" + + "\021sample_rate_hertz\030\002 \001(\005\022T\n\030synthesize_s" + + "peech_config\030\003 \001(\01322.google.cloud.dialog" + + "flow.v2.SynthesizeSpeechConfig*\373\001\n\rAudio" + + "Encoding\022\036\n\032AUDIO_ENCODING_UNSPECIFIED\020\000" + + "\022\034\n\030AUDIO_ENCODING_LINEAR_16\020\001\022\027\n\023AUDIO_" + + "ENCODING_FLAC\020\002\022\030\n\024AUDIO_ENCODING_MULAW\020" + + "\003\022\026\n\022AUDIO_ENCODING_AMR\020\004\022\031\n\025AUDIO_ENCOD" + + "ING_AMR_WB\020\005\022\033\n\027AUDIO_ENCODING_OGG_OPUS\020" + + "\006\022)\n%AUDIO_ENCODING_SPEEX_WITH_HEADER_BY" + + "TE\020\007*v\n\022SpeechModelVariant\022$\n SPEECH_MOD" + + "EL_VARIANT_UNSPECIFIED\020\000\022\026\n\022USE_BEST_AVA" + + "ILABLE\020\001\022\020\n\014USE_STANDARD\020\002\022\020\n\014USE_ENHANC" + + "ED\020\003*\215\001\n\017SsmlVoiceGender\022!\n\035SSML_VOICE_G" + + "ENDER_UNSPECIFIED\020\000\022\032\n\026SSML_VOICE_GENDER" + + "_MALE\020\001\022\034\n\030SSML_VOICE_GENDER_FEMALE\020\002\022\035\n" + + "\031SSML_VOICE_GENDER_NEUTRAL\020\003*\244\001\n\023OutputA" + + "udioEncoding\022%\n!OUTPUT_AUDIO_ENCODING_UN" + + "SPECIFIED\020\000\022#\n\037OUTPUT_AUDIO_ENCODING_LIN" + + "EAR_16\020\001\022\035\n\031OUTPUT_AUDIO_ENCODING_MP3\020\002\022" + + "\"\n\036OUTPUT_AUDIO_ENCODING_OGG_OPUS\020\003B\237\001\n\036" + + "com.google.cloud.dialogflow.v2B\020AudioCon" + + "figProtoP\001ZDgoogle.golang.org/genproto/g" + + "oogleapis/cloud/dialogflow/v2;dialogflow" + + "\370\001\001\242\002\002DF\252\002\032Google.Cloud.Dialogflow.V2b\006p" + + "roto3" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { @@ -75,8 +95,16 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors( com.google.api.AnnotationsProto.getDescriptor(), }, assigner); - internal_static_google_cloud_dialogflow_v2_VoiceSelectionParams_descriptor = + internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor = getDescriptor().getMessageTypes().get(0); + internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable = + new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( + internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor, + new java.lang.String[] { + "AudioEncoding", "SampleRateHertz", "LanguageCode", "PhraseHints", "ModelVariant", + }); + internal_static_google_cloud_dialogflow_v2_VoiceSelectionParams_descriptor = + getDescriptor().getMessageTypes().get(1); internal_static_google_cloud_dialogflow_v2_VoiceSelectionParams_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dialogflow_v2_VoiceSelectionParams_descriptor, @@ -84,7 +112,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors( "Name", "SsmlGender", }); internal_static_google_cloud_dialogflow_v2_SynthesizeSpeechConfig_descriptor = - getDescriptor().getMessageTypes().get(1); + getDescriptor().getMessageTypes().get(2); internal_static_google_cloud_dialogflow_v2_SynthesizeSpeechConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dialogflow_v2_SynthesizeSpeechConfig_descriptor, @@ -92,7 +120,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors( "SpeakingRate", "Pitch", "VolumeGainDb", "EffectsProfileId", "Voice", }); internal_static_google_cloud_dialogflow_v2_OutputAudioConfig_descriptor = - getDescriptor().getMessageTypes().get(2); + getDescriptor().getMessageTypes().get(3); internal_static_google_cloud_dialogflow_v2_OutputAudioConfig_fieldAccessorTable = new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_google_cloud_dialogflow_v2_OutputAudioConfig_descriptor, diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioEncoding.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioEncoding.java index 29d255466510..abf8ded84550 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioEncoding.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/AudioEncoding.java @@ -1,5 +1,5 @@ // Generated by the protocol buffer compiler. DO NOT EDIT! -// source: google/cloud/dialogflow/v2/session.proto +// source: google/cloud/dialogflow/v2/audio_config.proto package com.google.cloud.dialogflow.v2; @@ -274,7 +274,7 @@ public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType } public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { - return com.google.cloud.dialogflow.v2.SessionProto.getDescriptor().getEnumTypes().get(0); + return com.google.cloud.dialogflow.v2.AudioConfigProto.getDescriptor().getEnumTypes().get(0); } private static final AudioEncoding[] VALUES = values(); diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfig.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfig.java index c361c911cb1e..67a572c50889 100644 --- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfig.java +++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfig.java @@ -1,5 +1,5 @@ // Generated by the protocol buffer compiler. DO NOT EDIT! -// source: google/cloud/dialogflow/v2/session.proto +// source: google/cloud/dialogflow/v2/audio_config.proto package com.google.cloud.dialogflow.v2; @@ -26,6 +26,7 @@ private InputAudioConfig() { audioEncoding_ = 0; languageCode_ = ""; phraseHints_ = com.google.protobuf.LazyStringArrayList.EMPTY; + modelVariant_ = 0; } @java.lang.Override @@ -81,6 +82,13 @@ private InputAudioConfig( phraseHints_.add(s); break; } + case 80: + { + int rawValue = input.readEnum(); + + modelVariant_ = rawValue; + break; + } default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { @@ -104,14 +112,14 @@ private InputAudioConfig( } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { - return com.google.cloud.dialogflow.v2.SessionProto + return com.google.cloud.dialogflow.v2.AudioConfigProto .internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor; } @java.lang.Override protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { - return com.google.cloud.dialogflow.v2.SessionProto + return com.google.cloud.dialogflow.v2.AudioConfigProto .internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable .ensureFieldAccessorsInitialized( com.google.cloud.dialogflow.v2.InputAudioConfig.class, @@ -290,6 +298,36 @@ public com.google.protobuf.ByteString getPhraseHintsBytes(int index) { return phraseHints_.getByteString(index); } + public static final int MODEL_VARIANT_FIELD_NUMBER = 10; + private int modelVariant_; + /** + * + * + *
+ * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10;
+ */
+ public int getModelVariantValue() {
+ return modelVariant_;
+ }
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10;
+ */
+ public com.google.cloud.dialogflow.v2.SpeechModelVariant getModelVariant() {
+ @SuppressWarnings("deprecation")
+ com.google.cloud.dialogflow.v2.SpeechModelVariant result =
+ com.google.cloud.dialogflow.v2.SpeechModelVariant.valueOf(modelVariant_);
+ return result == null ? com.google.cloud.dialogflow.v2.SpeechModelVariant.UNRECOGNIZED : result;
+ }
+
private byte memoizedIsInitialized = -1;
@java.lang.Override
@@ -317,6 +355,11 @@ public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io
for (int i = 0; i < phraseHints_.size(); i++) {
com.google.protobuf.GeneratedMessageV3.writeString(output, 4, phraseHints_.getRaw(i));
}
+ if (modelVariant_
+ != com.google.cloud.dialogflow.v2.SpeechModelVariant.SPEECH_MODEL_VARIANT_UNSPECIFIED
+ .getNumber()) {
+ output.writeEnum(10, modelVariant_);
+ }
unknownFields.writeTo(output);
}
@@ -344,6 +387,11 @@ public int getSerializedSize() {
size += dataSize;
size += 1 * getPhraseHintsList().size();
}
+ if (modelVariant_
+ != com.google.cloud.dialogflow.v2.SpeechModelVariant.SPEECH_MODEL_VARIANT_UNSPECIFIED
+ .getNumber()) {
+ size += com.google.protobuf.CodedOutputStream.computeEnumSize(10, modelVariant_);
+ }
size += unknownFields.getSerializedSize();
memoizedSize = size;
return size;
@@ -364,6 +412,7 @@ public boolean equals(final java.lang.Object obj) {
if (getSampleRateHertz() != other.getSampleRateHertz()) return false;
if (!getLanguageCode().equals(other.getLanguageCode())) return false;
if (!getPhraseHintsList().equals(other.getPhraseHintsList())) return false;
+ if (modelVariant_ != other.modelVariant_) return false;
if (!unknownFields.equals(other.unknownFields)) return false;
return true;
}
@@ -385,6 +434,8 @@ public int hashCode() {
hash = (37 * hash) + PHRASE_HINTS_FIELD_NUMBER;
hash = (53 * hash) + getPhraseHintsList().hashCode();
}
+ hash = (37 * hash) + MODEL_VARIANT_FIELD_NUMBER;
+ hash = (53 * hash) + modelVariant_;
hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
@@ -499,14 +550,14 @@ public static final class Builder extends com.google.protobuf.GeneratedMessageV3
// @@protoc_insertion_point(builder_implements:google.cloud.dialogflow.v2.InputAudioConfig)
com.google.cloud.dialogflow.v2.InputAudioConfigOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() {
- return com.google.cloud.dialogflow.v2.SessionProto
+ return com.google.cloud.dialogflow.v2.AudioConfigProto
.internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor;
}
@java.lang.Override
protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
- return com.google.cloud.dialogflow.v2.SessionProto
+ return com.google.cloud.dialogflow.v2.AudioConfigProto
.internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable
.ensureFieldAccessorsInitialized(
com.google.cloud.dialogflow.v2.InputAudioConfig.class,
@@ -538,12 +589,14 @@ public Builder clear() {
phraseHints_ = com.google.protobuf.LazyStringArrayList.EMPTY;
bitField0_ = (bitField0_ & ~0x00000008);
+ modelVariant_ = 0;
+
return this;
}
@java.lang.Override
public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() {
- return com.google.cloud.dialogflow.v2.SessionProto
+ return com.google.cloud.dialogflow.v2.AudioConfigProto
.internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor;
}
@@ -575,6 +628,7 @@ public com.google.cloud.dialogflow.v2.InputAudioConfig buildPartial() {
bitField0_ = (bitField0_ & ~0x00000008);
}
result.phraseHints_ = phraseHints_;
+ result.modelVariant_ = modelVariant_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
@@ -646,6 +700,9 @@ public Builder mergeFrom(com.google.cloud.dialogflow.v2.InputAudioConfig other)
}
onChanged();
}
+ if (other.modelVariant_ != 0) {
+ setModelVariantValue(other.getModelVariantValue());
+ }
this.mergeUnknownFields(other.unknownFields);
onChanged();
return this;
@@ -1111,6 +1168,84 @@ public Builder addPhraseHintsBytes(com.google.protobuf.ByteString value) {
return this;
}
+ private int modelVariant_ = 0;
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10;
+ */
+ public int getModelVariantValue() {
+ return modelVariant_;
+ }
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10;
+ */
+ public Builder setModelVariantValue(int value) {
+ modelVariant_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10;
+ */
+ public com.google.cloud.dialogflow.v2.SpeechModelVariant getModelVariant() {
+ @SuppressWarnings("deprecation")
+ com.google.cloud.dialogflow.v2.SpeechModelVariant result =
+ com.google.cloud.dialogflow.v2.SpeechModelVariant.valueOf(modelVariant_);
+ return result == null
+ ? com.google.cloud.dialogflow.v2.SpeechModelVariant.UNRECOGNIZED
+ : result;
+ }
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10;
+ */
+ public Builder setModelVariant(com.google.cloud.dialogflow.v2.SpeechModelVariant value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+
+ modelVariant_ = value.getNumber();
+ onChanged();
+ return this;
+ }
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10;
+ */
+ public Builder clearModelVariant() {
+
+ modelVariant_ = 0;
+ onChanged();
+ return this;
+ }
+
@java.lang.Override
public final Builder setUnknownFields(final com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfigOrBuilder.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfigOrBuilder.java
index f1b0da5bf5fc..c8d0655150a7 100644
--- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfigOrBuilder.java
+++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/InputAudioConfigOrBuilder.java
@@ -1,5 +1,5 @@
// Generated by the protocol buffer compiler. DO NOT EDIT!
-// source: google/cloud/dialogflow/v2/session.proto
+// source: google/cloud/dialogflow/v2/audio_config.proto
package com.google.cloud.dialogflow.v2;
@@ -133,4 +133,25 @@ public interface InputAudioConfigOrBuilder
* repeated string phrase_hints = 4;
*/
com.google.protobuf.ByteString getPhraseHintsBytes(int index);
+
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10;
+ */
+ int getModelVariantValue();
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2.SpeechModelVariant model_variant = 10;
+ */
+ com.google.cloud.dialogflow.v2.SpeechModelVariant getModelVariant();
}
diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioConfig.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioConfig.java
index 3f0a1b4890e2..16b8c21fa7d7 100644
--- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioConfig.java
+++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioConfig.java
@@ -7,7 +7,7 @@
*
*
* - * Instructs the speech synthesizer how to generate the output audio content. + * Instructs the speech synthesizer on how to generate the output audio content. ** * Protobuf type {@code google.cloud.dialogflow.v2.OutputAudioConfig} @@ -397,7 +397,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build * * *
- * Instructs the speech synthesizer how to generate the output audio content.
+ * Instructs the speech synthesizer on how to generate the output audio content.
*
*
* Protobuf type {@code google.cloud.dialogflow.v2.OutputAudioConfig}
diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioEncoding.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioEncoding.java
index c03b8e0ba80e..16c0392a621b 100644
--- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioEncoding.java
+++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/OutputAudioEncoding.java
@@ -156,7 +156,7 @@ public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType
}
public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() {
- return com.google.cloud.dialogflow.v2.AudioConfigProto.getDescriptor().getEnumTypes().get(1);
+ return com.google.cloud.dialogflow.v2.AudioConfigProto.getDescriptor().getEnumTypes().get(3);
}
private static final OutputAudioEncoding[] VALUES = values();
diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SessionProto.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SessionProto.java
index b93b98ec2e0b..c579791b0940 100644
--- a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SessionProto.java
+++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SessionProto.java
@@ -44,10 +44,6 @@ public static void registerAllExtensions(com.google.protobuf.ExtensionRegistry r
internal_static_google_cloud_dialogflow_v2_StreamingRecognitionResult_descriptor;
static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_google_cloud_dialogflow_v2_StreamingRecognitionResult_fieldAccessorTable;
- static final com.google.protobuf.Descriptors.Descriptor
- internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor;
- static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
- internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable;
static final com.google.protobuf.Descriptors.Descriptor
internal_static_google_cloud_dialogflow_v2_TextInput_descriptor;
static final com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
@@ -153,11 +149,7 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() {
+ "ype\022\022\n\ntranscript\030\002 \001(\t\022\020\n\010is_final\030\003 \001("
+ "\010\022\022\n\nconfidence\030\004 \001(\002\"X\n\013MessageType\022\034\n\030"
+ "MESSAGE_TYPE_UNSPECIFIED\020\000\022\016\n\nTRANSCRIPT"
- + "\020\001\022\033\n\027END_OF_SINGLE_UTTERANCE\020\002\"\235\001\n\020Inpu"
- + "tAudioConfig\022A\n\016audio_encoding\030\001 \001(\0162).g"
- + "oogle.cloud.dialogflow.v2.AudioEncoding\022"
- + "\031\n\021sample_rate_hertz\030\002 \001(\005\022\025\n\rlanguage_c"
- + "ode\030\003 \001(\t\022\024\n\014phrase_hints\030\004 \003(\t\"0\n\tTextI"
+ + "\020\001\022\033\n\027END_OF_SINGLE_UTTERANCE\020\002\"0\n\tTextI"
+ "nput\022\014\n\004text\030\001 \001(\t\022\025\n\rlanguage_code\030\002 \001("
+ "\t\"^\n\nEventInput\022\014\n\004name\030\001 \001(\t\022+\n\nparamet"
+ "ers\030\002 \001(\0132\027.google.protobuf.Struct\022\025\n\rla"
@@ -166,26 +158,20 @@ public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() {
+ "ent\030\001 \001(\010\"^\n\027SentimentAnalysisResult\022C\n\024"
+ "query_text_sentiment\030\001 \001(\0132%.google.clou"
+ "d.dialogflow.v2.Sentiment\"-\n\tSentiment\022\r"
- + "\n\005score\030\001 \001(\002\022\021\n\tmagnitude\030\002 \001(\002*\373\001\n\rAud"
- + "ioEncoding\022\036\n\032AUDIO_ENCODING_UNSPECIFIED"
- + "\020\000\022\034\n\030AUDIO_ENCODING_LINEAR_16\020\001\022\027\n\023AUDI"
- + "O_ENCODING_FLAC\020\002\022\030\n\024AUDIO_ENCODING_MULA"
- + "W\020\003\022\026\n\022AUDIO_ENCODING_AMR\020\004\022\031\n\025AUDIO_ENC"
- + "ODING_AMR_WB\020\005\022\033\n\027AUDIO_ENCODING_OGG_OPU"
- + "S\020\006\022)\n%AUDIO_ENCODING_SPEEX_WITH_HEADER_"
- + "BYTE\020\0072\326\002\n\010Sessions\022\264\001\n\014DetectIntent\022/.g"
- + "oogle.cloud.dialogflow.v2.DetectIntentRe"
- + "quest\0320.google.cloud.dialogflow.v2.Detec"
- + "tIntentResponse\"A\202\323\344\223\002;\"6/v2/{session=pr"
- + "ojects/*/agent/sessions/*}:detectIntent:"
- + "\001*\022\222\001\n\025StreamingDetectIntent\0228.google.cl"
- + "oud.dialogflow.v2.StreamingDetectIntentR"
- + "equest\0329.google.cloud.dialogflow.v2.Stre"
- + "amingDetectIntentResponse\"\000(\0010\001B\233\001\n\036com."
- + "google.cloud.dialogflow.v2B\014SessionProto"
- + "P\001ZDgoogle.golang.org/genproto/googleapi"
- + "s/cloud/dialogflow/v2;dialogflow\370\001\001\242\002\002DF"
- + "\252\002\032Google.Cloud.Dialogflow.V2b\006proto3"
+ + "\n\005score\030\001 \001(\002\022\021\n\tmagnitude\030\002 \001(\0022\326\002\n\010Ses"
+ + "sions\022\264\001\n\014DetectIntent\022/.google.cloud.di"
+ + "alogflow.v2.DetectIntentRequest\0320.google"
+ + ".cloud.dialogflow.v2.DetectIntentRespons"
+ + "e\"A\202\323\344\223\002;\"6/v2/{session=projects/*/agent"
+ + "/sessions/*}:detectIntent:\001*\022\222\001\n\025Streami"
+ + "ngDetectIntent\0228.google.cloud.dialogflow"
+ + ".v2.StreamingDetectIntentRequest\0329.googl"
+ + "e.cloud.dialogflow.v2.StreamingDetectInt"
+ + "entResponse\"\000(\0010\001B\233\001\n\036com.google.cloud.d"
+ + "ialogflow.v2B\014SessionProtoP\001ZDgoogle.gol"
+ + "ang.org/genproto/googleapis/cloud/dialog"
+ + "flow/v2;dialogflow\370\001\001\242\002\002DF\252\002\032Google.Clou"
+ + "d.Dialogflow.V2b\006proto3"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
@@ -303,16 +289,8 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors(
new java.lang.String[] {
"MessageType", "Transcript", "IsFinal", "Confidence",
});
- internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor =
- getDescriptor().getMessageTypes().get(8);
- internal_static_google_cloud_dialogflow_v2_InputAudioConfig_fieldAccessorTable =
- new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
- internal_static_google_cloud_dialogflow_v2_InputAudioConfig_descriptor,
- new java.lang.String[] {
- "AudioEncoding", "SampleRateHertz", "LanguageCode", "PhraseHints",
- });
internal_static_google_cloud_dialogflow_v2_TextInput_descriptor =
- getDescriptor().getMessageTypes().get(9);
+ getDescriptor().getMessageTypes().get(8);
internal_static_google_cloud_dialogflow_v2_TextInput_fieldAccessorTable =
new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_google_cloud_dialogflow_v2_TextInput_descriptor,
@@ -320,7 +298,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors(
"Text", "LanguageCode",
});
internal_static_google_cloud_dialogflow_v2_EventInput_descriptor =
- getDescriptor().getMessageTypes().get(10);
+ getDescriptor().getMessageTypes().get(9);
internal_static_google_cloud_dialogflow_v2_EventInput_fieldAccessorTable =
new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_google_cloud_dialogflow_v2_EventInput_descriptor,
@@ -328,7 +306,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors(
"Name", "Parameters", "LanguageCode",
});
internal_static_google_cloud_dialogflow_v2_SentimentAnalysisRequestConfig_descriptor =
- getDescriptor().getMessageTypes().get(11);
+ getDescriptor().getMessageTypes().get(10);
internal_static_google_cloud_dialogflow_v2_SentimentAnalysisRequestConfig_fieldAccessorTable =
new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_google_cloud_dialogflow_v2_SentimentAnalysisRequestConfig_descriptor,
@@ -336,7 +314,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors(
"AnalyzeQueryTextSentiment",
});
internal_static_google_cloud_dialogflow_v2_SentimentAnalysisResult_descriptor =
- getDescriptor().getMessageTypes().get(12);
+ getDescriptor().getMessageTypes().get(11);
internal_static_google_cloud_dialogflow_v2_SentimentAnalysisResult_fieldAccessorTable =
new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_google_cloud_dialogflow_v2_SentimentAnalysisResult_descriptor,
@@ -344,7 +322,7 @@ public com.google.protobuf.ExtensionRegistry assignDescriptors(
"QueryTextSentiment",
});
internal_static_google_cloud_dialogflow_v2_Sentiment_descriptor =
- getDescriptor().getMessageTypes().get(13);
+ getDescriptor().getMessageTypes().get(12);
internal_static_google_cloud_dialogflow_v2_Sentiment_fieldAccessorTable =
new com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_google_cloud_dialogflow_v2_Sentiment_descriptor,
diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SpeechModelVariant.java b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SpeechModelVariant.java
new file mode 100644
index 000000000000..ffaf71bf0ecc
--- /dev/null
+++ b/google-api-grpc/proto-google-cloud-dialogflow-v2/src/main/java/com/google/cloud/dialogflow/v2/SpeechModelVariant.java
@@ -0,0 +1,214 @@
+// Generated by the protocol buffer compiler. DO NOT EDIT!
+// source: google/cloud/dialogflow/v2/audio_config.proto
+
+package com.google.cloud.dialogflow.v2;
+
+/**
+ *
+ *
+ * + * Variant of the specified [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use. + * See the [Cloud Speech + * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + * for which models have different variants. For example, the "phone_call" model + * has both a standard and an enhanced variant. When you use an enhanced model, + * you will generally receive higher quality results than for a standard model. + *+ * + * Protobuf enum {@code google.cloud.dialogflow.v2.SpeechModelVariant} + */ +public enum SpeechModelVariant implements com.google.protobuf.ProtocolMessageEnum { + /** + * + * + *
+ * No model variant specified. In this case Dialogflow defaults to + * USE_BEST_AVAILABLE. + *+ * + *
SPEECH_MODEL_VARIANT_UNSPECIFIED = 0;
+ */
+ SPEECH_MODEL_VARIANT_UNSPECIFIED(0),
+ /**
+ *
+ *
+ * + * Use the best available variant of the [Speech + * model][InputAudioConfig.model] that the caller is eligible for. + * Please see the [Dialogflow + * docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) for + * how to make your project eligible for enhanced models. + *+ * + *
USE_BEST_AVAILABLE = 1;
+ */
+ USE_BEST_AVAILABLE(1),
+ /**
+ *
+ *
+ * + * Use standard model variant even if an enhanced model is available. See the + * [Cloud Speech + * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + * for details about enhanced models. + *+ * + *
USE_STANDARD = 2;
+ */
+ USE_STANDARD(2),
+ /**
+ *
+ *
+ * + * Use an enhanced model variant: + * * If an enhanced variant does not exist for the given + * [model][google.cloud.dialogflow.v2.InputAudioConfig.model] and request language, Dialogflow falls + * back to the standard variant. + * The [Cloud Speech + * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + * describes which models have enhanced variants. + * * If the API caller isn't eligible for enhanced models, Dialogflow returns + * an error. Please see the [Dialogflow + * docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) + * for how to make your project eligible. + *+ * + *
USE_ENHANCED = 3;
+ */
+ USE_ENHANCED(3),
+ UNRECOGNIZED(-1),
+ ;
+
+ /**
+ *
+ *
+ * + * No model variant specified. In this case Dialogflow defaults to + * USE_BEST_AVAILABLE. + *+ * + *
SPEECH_MODEL_VARIANT_UNSPECIFIED = 0;
+ */
+ public static final int SPEECH_MODEL_VARIANT_UNSPECIFIED_VALUE = 0;
+ /**
+ *
+ *
+ * + * Use the best available variant of the [Speech + * model][InputAudioConfig.model] that the caller is eligible for. + * Please see the [Dialogflow + * docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) for + * how to make your project eligible for enhanced models. + *+ * + *
USE_BEST_AVAILABLE = 1;
+ */
+ public static final int USE_BEST_AVAILABLE_VALUE = 1;
+ /**
+ *
+ *
+ * + * Use standard model variant even if an enhanced model is available. See the + * [Cloud Speech + * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + * for details about enhanced models. + *+ * + *
USE_STANDARD = 2;
+ */
+ public static final int USE_STANDARD_VALUE = 2;
+ /**
+ *
+ *
+ * + * Use an enhanced model variant: + * * If an enhanced variant does not exist for the given + * [model][google.cloud.dialogflow.v2.InputAudioConfig.model] and request language, Dialogflow falls + * back to the standard variant. + * The [Cloud Speech + * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + * describes which models have enhanced variants. + * * If the API caller isn't eligible for enhanced models, Dialogflow returns + * an error. Please see the [Dialogflow + * docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) + * for how to make your project eligible. + *+ * + *
USE_ENHANCED = 3;
+ */
+ public static final int USE_ENHANCED_VALUE = 3;
+
+ public final int getNumber() {
+ if (this == UNRECOGNIZED) {
+ throw new java.lang.IllegalArgumentException(
+ "Can't get the number of an unknown enum value.");
+ }
+ return value;
+ }
+
+ /** @deprecated Use {@link #forNumber(int)} instead. */
+ @java.lang.Deprecated
+ public static SpeechModelVariant valueOf(int value) {
+ return forNumber(value);
+ }
+
+ public static SpeechModelVariant forNumber(int value) {
+ switch (value) {
+ case 0:
+ return SPEECH_MODEL_VARIANT_UNSPECIFIED;
+ case 1:
+ return USE_BEST_AVAILABLE;
+ case 2:
+ return USE_STANDARD;
+ case 3:
+ return USE_ENHANCED;
+ default:
+ return null;
+ }
+ }
+
+ public static com.google.protobuf.Internal.EnumLiteMap- * Instructs the speech recognizer how to process the audio content. + * Instructs the speech recognizer on how to process the audio content. ** * Protobuf type {@code google.cloud.dialogflow.v2beta1.InputAudioConfig} @@ -27,6 +27,7 @@ private InputAudioConfig() { languageCode_ = ""; phraseHints_ = com.google.protobuf.LazyStringArrayList.EMPTY; model_ = ""; + modelVariant_ = 0; } @java.lang.Override @@ -89,6 +90,13 @@ private InputAudioConfig( model_ = s; break; } + case 80: + { + int rawValue = input.readEnum(); + + modelVariant_ = rawValue; + break; + } default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { @@ -361,6 +369,38 @@ public com.google.protobuf.ByteString getModelBytes() { } } + public static final int MODEL_VARIANT_FIELD_NUMBER = 10; + private int modelVariant_; + /** + * + * + *
+ * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10;
+ */
+ public int getModelVariantValue() {
+ return modelVariant_;
+ }
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10;
+ */
+ public com.google.cloud.dialogflow.v2beta1.SpeechModelVariant getModelVariant() {
+ @SuppressWarnings("deprecation")
+ com.google.cloud.dialogflow.v2beta1.SpeechModelVariant result =
+ com.google.cloud.dialogflow.v2beta1.SpeechModelVariant.valueOf(modelVariant_);
+ return result == null
+ ? com.google.cloud.dialogflow.v2beta1.SpeechModelVariant.UNRECOGNIZED
+ : result;
+ }
+
private byte memoizedIsInitialized = -1;
@java.lang.Override
@@ -392,6 +432,11 @@ public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io
if (!getModelBytes().isEmpty()) {
com.google.protobuf.GeneratedMessageV3.writeString(output, 7, model_);
}
+ if (modelVariant_
+ != com.google.cloud.dialogflow.v2beta1.SpeechModelVariant.SPEECH_MODEL_VARIANT_UNSPECIFIED
+ .getNumber()) {
+ output.writeEnum(10, modelVariant_);
+ }
unknownFields.writeTo(output);
}
@@ -423,6 +468,11 @@ public int getSerializedSize() {
if (!getModelBytes().isEmpty()) {
size += com.google.protobuf.GeneratedMessageV3.computeStringSize(7, model_);
}
+ if (modelVariant_
+ != com.google.cloud.dialogflow.v2beta1.SpeechModelVariant.SPEECH_MODEL_VARIANT_UNSPECIFIED
+ .getNumber()) {
+ size += com.google.protobuf.CodedOutputStream.computeEnumSize(10, modelVariant_);
+ }
size += unknownFields.getSerializedSize();
memoizedSize = size;
return size;
@@ -444,6 +494,7 @@ public boolean equals(final java.lang.Object obj) {
if (!getLanguageCode().equals(other.getLanguageCode())) return false;
if (!getPhraseHintsList().equals(other.getPhraseHintsList())) return false;
if (!getModel().equals(other.getModel())) return false;
+ if (modelVariant_ != other.modelVariant_) return false;
if (!unknownFields.equals(other.unknownFields)) return false;
return true;
}
@@ -467,6 +518,8 @@ public int hashCode() {
}
hash = (37 * hash) + MODEL_FIELD_NUMBER;
hash = (53 * hash) + getModel().hashCode();
+ hash = (37 * hash) + MODEL_VARIANT_FIELD_NUMBER;
+ hash = (53 * hash) + modelVariant_;
hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
@@ -571,7 +624,7 @@ protected Builder newBuilderForType(com.google.protobuf.GeneratedMessageV3.Build
*
*
*
- * Instructs the speech recognizer how to process the audio content.
+ * Instructs the speech recognizer on how to process the audio content.
*
*
* Protobuf type {@code google.cloud.dialogflow.v2beta1.InputAudioConfig}
@@ -622,6 +675,8 @@ public Builder clear() {
bitField0_ = (bitField0_ & ~0x00000008);
model_ = "";
+ modelVariant_ = 0;
+
return this;
}
@@ -660,6 +715,7 @@ public com.google.cloud.dialogflow.v2beta1.InputAudioConfig buildPartial() {
}
result.phraseHints_ = phraseHints_;
result.model_ = model_;
+ result.modelVariant_ = modelVariant_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
@@ -735,6 +791,9 @@ public Builder mergeFrom(com.google.cloud.dialogflow.v2beta1.InputAudioConfig ot
model_ = other.model_;
onChanged();
}
+ if (other.modelVariant_ != 0) {
+ setModelVariantValue(other.getModelVariantValue());
+ }
this.mergeUnknownFields(other.unknownFields);
onChanged();
return this;
@@ -1347,6 +1406,84 @@ public Builder setModelBytes(com.google.protobuf.ByteString value) {
return this;
}
+ private int modelVariant_ = 0;
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10;
+ */
+ public int getModelVariantValue() {
+ return modelVariant_;
+ }
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10;
+ */
+ public Builder setModelVariantValue(int value) {
+ modelVariant_ = value;
+ onChanged();
+ return this;
+ }
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10;
+ */
+ public com.google.cloud.dialogflow.v2beta1.SpeechModelVariant getModelVariant() {
+ @SuppressWarnings("deprecation")
+ com.google.cloud.dialogflow.v2beta1.SpeechModelVariant result =
+ com.google.cloud.dialogflow.v2beta1.SpeechModelVariant.valueOf(modelVariant_);
+ return result == null
+ ? com.google.cloud.dialogflow.v2beta1.SpeechModelVariant.UNRECOGNIZED
+ : result;
+ }
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10;
+ */
+ public Builder setModelVariant(com.google.cloud.dialogflow.v2beta1.SpeechModelVariant value) {
+ if (value == null) {
+ throw new NullPointerException();
+ }
+
+ modelVariant_ = value.getNumber();
+ onChanged();
+ return this;
+ }
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10;
+ */
+ public Builder clearModelVariant() {
+
+ modelVariant_ = 0;
+ onChanged();
+ return this;
+ }
+
@java.lang.Override
public final Builder setUnknownFields(final com.google.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/InputAudioConfigOrBuilder.java b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/InputAudioConfigOrBuilder.java
index a2a2c48b8644..da25c3ef3063 100644
--- a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/InputAudioConfigOrBuilder.java
+++ b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/InputAudioConfigOrBuilder.java
@@ -174,4 +174,25 @@ public interface InputAudioConfigOrBuilder
* string model = 7;
*/
com.google.protobuf.ByteString getModelBytes();
+
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10;
+ */
+ int getModelVariantValue();
+ /**
+ *
+ *
+ * + * Optional. Which variant of the [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use. + *+ * + *
.google.cloud.dialogflow.v2beta1.SpeechModelVariant model_variant = 10;
+ */
+ com.google.cloud.dialogflow.v2beta1.SpeechModelVariant getModelVariant();
}
diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/OutputAudioEncoding.java b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/OutputAudioEncoding.java
index f46dee66703a..1de522c1c579 100644
--- a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/OutputAudioEncoding.java
+++ b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/OutputAudioEncoding.java
@@ -158,7 +158,7 @@ public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType
public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() {
return com.google.cloud.dialogflow.v2beta1.AudioConfigProto.getDescriptor()
.getEnumTypes()
- .get(2);
+ .get(3);
}
private static final OutputAudioEncoding[] VALUES = values();
diff --git a/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/SpeechModelVariant.java b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/SpeechModelVariant.java
new file mode 100644
index 000000000000..c08c12c25e7c
--- /dev/null
+++ b/google-api-grpc/proto-google-cloud-dialogflow-v2beta1/src/main/java/com/google/cloud/dialogflow/v2beta1/SpeechModelVariant.java
@@ -0,0 +1,216 @@
+// Generated by the protocol buffer compiler. DO NOT EDIT!
+// source: google/cloud/dialogflow/v2beta1/audio_config.proto
+
+package com.google.cloud.dialogflow.v2beta1;
+
+/**
+ *
+ *
+ * + * Variant of the specified [Speech model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] to use. + * See the [Cloud Speech + * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + * for which models have different variants. For example, the "phone_call" model + * has both a standard and an enhanced variant. When you use an enhanced model, + * you will generally receive higher quality results than for a standard model. + *+ * + * Protobuf enum {@code google.cloud.dialogflow.v2beta1.SpeechModelVariant} + */ +public enum SpeechModelVariant implements com.google.protobuf.ProtocolMessageEnum { + /** + * + * + *
+ * No model variant specified. In this case Dialogflow defaults to + * USE_BEST_AVAILABLE. + *+ * + *
SPEECH_MODEL_VARIANT_UNSPECIFIED = 0;
+ */
+ SPEECH_MODEL_VARIANT_UNSPECIFIED(0),
+ /**
+ *
+ *
+ * + * Use the best available variant of the [Speech + * model][InputAudioConfig.model] that the caller is eligible for. + * Please see the [Dialogflow + * docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) for + * how to make your project eligible for enhanced models. + *+ * + *
USE_BEST_AVAILABLE = 1;
+ */
+ USE_BEST_AVAILABLE(1),
+ /**
+ *
+ *
+ * + * Use standard model variant even if an enhanced model is available. See the + * [Cloud Speech + * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + * for details about enhanced models. + *+ * + *
USE_STANDARD = 2;
+ */
+ USE_STANDARD(2),
+ /**
+ *
+ *
+ * + * Use an enhanced model variant: + * * If an enhanced variant does not exist for the given + * [model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] and request language, Dialogflow falls + * back to the standard variant. + * The [Cloud Speech + * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + * describes which models have enhanced variants. + * * If the API caller isn't eligible for enhanced models, Dialogflow returns + * an error. Please see the [Dialogflow + * docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) + * for how to make your project eligible. + *+ * + *
USE_ENHANCED = 3;
+ */
+ USE_ENHANCED(3),
+ UNRECOGNIZED(-1),
+ ;
+
+ /**
+ *
+ *
+ * + * No model variant specified. In this case Dialogflow defaults to + * USE_BEST_AVAILABLE. + *+ * + *
SPEECH_MODEL_VARIANT_UNSPECIFIED = 0;
+ */
+ public static final int SPEECH_MODEL_VARIANT_UNSPECIFIED_VALUE = 0;
+ /**
+ *
+ *
+ * + * Use the best available variant of the [Speech + * model][InputAudioConfig.model] that the caller is eligible for. + * Please see the [Dialogflow + * docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) for + * how to make your project eligible for enhanced models. + *+ * + *
USE_BEST_AVAILABLE = 1;
+ */
+ public static final int USE_BEST_AVAILABLE_VALUE = 1;
+ /**
+ *
+ *
+ * + * Use standard model variant even if an enhanced model is available. See the + * [Cloud Speech + * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + * for details about enhanced models. + *+ * + *
USE_STANDARD = 2;
+ */
+ public static final int USE_STANDARD_VALUE = 2;
+ /**
+ *
+ *
+ * + * Use an enhanced model variant: + * * If an enhanced variant does not exist for the given + * [model][google.cloud.dialogflow.v2beta1.InputAudioConfig.model] and request language, Dialogflow falls + * back to the standard variant. + * The [Cloud Speech + * documentation](https://cloud.google.com/speech-to-text/docs/enhanced-models) + * describes which models have enhanced variants. + * * If the API caller isn't eligible for enhanced models, Dialogflow returns + * an error. Please see the [Dialogflow + * docs](https://cloud.google.com/dialogflow-enterprise/docs/data-logging) + * for how to make your project eligible. + *+ * + *
USE_ENHANCED = 3;
+ */
+ public static final int USE_ENHANCED_VALUE = 3;
+
+ public final int getNumber() {
+ if (this == UNRECOGNIZED) {
+ throw new java.lang.IllegalArgumentException(
+ "Can't get the number of an unknown enum value.");
+ }
+ return value;
+ }
+
+ /** @deprecated Use {@link #forNumber(int)} instead. */
+ @java.lang.Deprecated
+ public static SpeechModelVariant valueOf(int value) {
+ return forNumber(value);
+ }
+
+ public static SpeechModelVariant forNumber(int value) {
+ switch (value) {
+ case 0:
+ return SPEECH_MODEL_VARIANT_UNSPECIFIED;
+ case 1:
+ return USE_BEST_AVAILABLE;
+ case 2:
+ return USE_STANDARD;
+ case 3:
+ return USE_ENHANCED;
+ default:
+ return null;
+ }
+ }
+
+ public static com.google.protobuf.Internal.EnumLiteMap