Package types (2.8.1)

API documentation for dialogflow_v2.types package.

Classes

Agent

A Dialogflow agent is a virtual agent that handles conversations with your end-users. It is a natural language understanding module that understands the nuances of human language. Dialogflow translates end-user text or audio during a conversation to structured data that your apps and services can understand. You design and build a Dialogflow agent to handle the types of conversations required for your system.

For more information about agents, see the Agent guide <https://cloud.google.com/dialogflow/docs/agents-overview>__.

AgentAssistantFeedback

Detail feedback of Agent Assist result. .. attribute:: answer_relevance

Optional. Whether or not the suggested answer is relevant.

For example:

AgentAssistantRecord

Represents a record of a human agent assist answer. .. attribute:: article_suggestion_answer

Output only. The article suggestion answer.

:type: google.cloud.dialogflow_v2.types.ArticleAnswer

AnalyzeContentRequest

The request message for Participants.AnalyzeContent.

AnalyzeContentResponse

The response message for Participants.AnalyzeContent.

AnnotatedMessagePart

Represents a part of a message possibly annotated with an entity. The part can be an entity or purely a part of the message between two entities or message start/end.

AnswerFeedback

Represents feedback the customer has about the quality & correctness of a certain answer in a conversation.

AnswerRecord

Answer records are records to manage answer history and feedbacks for Dialogflow.

Currently, answer record includes:

  • human agent assistant article suggestion
  • human agent assistant faq article

It doesn't include:

  • DetectIntent intent matching
  • DetectIntent knowledge

Answer records are not related to the conversation history in the Dialogflow Console. A Record is generated even when the end-user disables conversation history in the console. Records are created when there's a human agent assistant suggestion generated.

A typical workflow for customers provide feedback to an answer is:

  1. For human agent assistant, customers get suggestion via ListSuggestions API. Together with the answers, AnswerRecord.name are returned to the customers.
  2. The customer uses the AnswerRecord.name to call the [UpdateAnswerRecord][] method to send feedback about a specific answer that they believe is wrong.

ArticleAnswer

Represents article answer. .. attribute:: title

The article title.

:type: str

AudioEncoding

Audio encoding of the audio content sent in the conversational query request. Refer to the Cloud Speech API documentation <https://cloud.google.com/speech-to-text/docs/basics>__ for more details.

AutomatedAgentConfig

Defines the Automated Agent to connect to a conversation. .. attribute:: agent

Required. ID of the Dialogflow agent environment to use.

This project needs to either be the same project as the conversation or you need to grant service-<Conversation Project Number>@gcp-sa-dialogflow.iam.gserviceaccount.com the Dialogflow API Service Agent role in this project.

Format: projects/<Project ID>/locations/<Location ID>/agent/environments/<Environment ID or '-'>. If environment is not specified, the default draft environment is used. Refer to DetectIntentRequest </dialogflow/docs/reference/rpc/google.cloud.dialogflow.v2#google.cloud.dialogflow.v2.DetectIntentRequest>__ for more details.

:type: str

AutomatedAgentReply

Represents a response from an automated agent. .. attribute:: detect_intent_response

Response of the Dialogflow Sessions.DetectIntent call.

:type: google.cloud.dialogflow_v2.types.DetectIntentResponse

BatchCreateEntitiesRequest

The request message for EntityTypes.BatchCreateEntities.

BatchDeleteEntitiesRequest

The request message for EntityTypes.BatchDeleteEntities.

BatchDeleteEntityTypesRequest

The request message for EntityTypes.BatchDeleteEntityTypes.

BatchDeleteIntentsRequest

The request message for Intents.BatchDeleteIntents.

BatchUpdateEntitiesRequest

The request message for EntityTypes.BatchUpdateEntities.

BatchUpdateEntityTypesRequest

The request message for EntityTypes.BatchUpdateEntityTypes.

BatchUpdateEntityTypesResponse

The response message for EntityTypes.BatchUpdateEntityTypes.

BatchUpdateIntentsRequest

BatchUpdateIntentsResponse

The response message for Intents.BatchUpdateIntents.

CompleteConversationRequest

The request message for Conversations.CompleteConversation.

Context

Dialogflow contexts are similar to natural language context. If a person says to you "they are orange", you need context in order to understand what "they" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent.

Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts.

For more information about context, see the Contexts guide <https://cloud.google.com/dialogflow/docs/contexts-overview>__.

Conversation

Represents a conversation. A conversation is an interaction between an agent, including live agents and Dialogflow agents, and a support customer. Conversations can include phone calls and text-based chat sessions.

ConversationEvent

Represents a notification sent to Pub/Sub subscribers for conversation lifecycle events.

ConversationPhoneNumber

Represents a phone number for telephony integration. It allows for connecting a particular conversation over telephony.

ConversationProfile

Defines the services to connect to incoming Dialogflow conversations.

CreateContextRequest

The request message for Contexts.CreateContext.

CreateConversationProfileRequest

The request message for ConversationProfiles.CreateConversationProfile.

CreateConversationRequest

The request message for Conversations.CreateConversation.

CreateDocumentRequest

Request message for Documents.CreateDocument.

CreateEntityTypeRequest

The request message for EntityTypes.CreateEntityType.

CreateEnvironmentRequest

The request message for Environments.CreateEnvironment.

CreateIntentRequest

The request message for Intents.CreateIntent.

CreateKnowledgeBaseRequest

Request message for KnowledgeBases.CreateKnowledgeBase.

CreateParticipantRequest

The request message for Participants.CreateParticipant.

CreateSessionEntityTypeRequest

The request message for SessionEntityTypes.CreateSessionEntityType.

CreateVersionRequest

The request message for Versions.CreateVersion.

DeleteAgentRequest

The request message for Agents.DeleteAgent.

DeleteAllContextsRequest

The request message for Contexts.DeleteAllContexts.

DeleteContextRequest

The request message for Contexts.DeleteContext.

DeleteConversationProfileRequest

The request message for ConversationProfiles.DeleteConversationProfile.

This operation fails if the conversation profile is still referenced from a phone number.

DeleteDocumentRequest

Request message for Documents.DeleteDocument.

DeleteEntityTypeRequest

The request message for EntityTypes.DeleteEntityType.

DeleteEnvironmentRequest

The request message for Environments.DeleteEnvironment.

DeleteIntentRequest

The request message for Intents.DeleteIntent.

DeleteKnowledgeBaseRequest

Request message for KnowledgeBases.DeleteKnowledgeBase.

DeleteSessionEntityTypeRequest

The request message for SessionEntityTypes.DeleteSessionEntityType.

DeleteVersionRequest

The request message for Versions.DeleteVersion.

DetectIntentRequest

The request to detect user's intent. .. attribute:: session

Required. The name of the session this query is sent to. Format: projects/<Project ID>/agent/sessions/<Session ID>, or projects/<Project ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>. If Environment ID is not specified, we assume default 'draft' environment (Environment ID might be referred to as environment name at some places). If User ID is not specified, we are using "-". It's up to the API caller to choose an appropriate Session ID and User Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of the Session ID and User ID must not exceed 36 characters.

For more information, see the API interactions guide <https://cloud.google.com/dialogflow/docs/api-overview>__.

Note: Always use agent versions for production traffic. See Versions and environments <https://cloud.google.com/dialogflow/es/docs/agents-versions>__.

:type: str

DetectIntentResponse

The message returned from the DetectIntent method. .. attribute:: response_id

The unique identifier of the response. It can be used to locate a response in the training example set or for reporting issues.

:type: str

Document

A knowledge document to be used by a KnowledgeBase.

For more information, see the knowledge base guide <https://cloud.google.com/dialogflow/docs/how/knowledge-bases>__.

Note: The projects.agent.knowledgeBases.documents resource is deprecated; only use projects.knowledgeBases.documents.

DtmfParameters

The message in the response that indicates the parameters of DTMF.

EntityType

Each intent parameter has a type, called the entity type, which dictates exactly how data from an end-user expression is extracted.

Dialogflow provides predefined system entities that can match many common types of data. For example, there are system entities for matching dates, times, colors, email addresses, and so on. You can also create your own custom entities for matching custom data. For example, you could define a vegetable entity that can match the types of vegetables available for purchase with a grocery store agent.

For more information, see the Entity guide <https://cloud.google.com/dialogflow/docs/entities-overview>__.

EntityTypeBatch

This message is a wrapper around a collection of entity types.

Environment

You can create multiple versions of your agent and publish them to separate environments.

When you edit an agent, you are editing the draft agent. At any point, you can save the draft agent as an agent version, which is an immutable snapshot of your agent.

When you save the draft agent, it is published to the default environment. When you create agent versions, you can publish them to custom environments. You can create a variety of custom environments for:

  • testing
  • development
  • production
  • etc.

For more information, see the versions and environments guide <https://cloud.google.com/dialogflow/docs/agents-versions>__.

EnvironmentHistory

The response message for Environments.GetEnvironmentHistory.

EventInput

Events allow for matching intents by event name instead of the natural language input. For instance, input <event: { name: "welcome_event", parameters: { name: "Sam" } }> can trigger a personalized welcome response. The parameter name may be used by the agent in the response: "Hello #welcome_event.name! What can I do for you today?".

ExportAgentRequest

The request message for Agents.ExportAgent.

ExportAgentResponse

The response message for Agents.ExportAgent.

FaqAnswer

Represents answer from "frequently asked questions". .. attribute:: answer

The piece of text from the source knowledge base document.

:type: str

Fulfillment

By default, your agent responds to a matched intent with a static response. As an alternative, you can provide a more dynamic response by using fulfillment. When you enable fulfillment for an intent, Dialogflow responds to that intent by calling a service that you define. For example, if an end-user wants to schedule a haircut on Friday, your service can check your database and respond to the end-user with availability information for Friday.

For more information, see the fulfillment guide <https://cloud.google.com/dialogflow/docs/fulfillment-overview>__.

GetAgentRequest

The request message for Agents.GetAgent.

GetContextRequest

The request message for Contexts.GetContext.

GetConversationProfileRequest

The request message for ConversationProfiles.GetConversationProfile.

GetConversationRequest

The request message for Conversations.GetConversation.

GetDocumentRequest

Request message for Documents.GetDocument.

GetEntityTypeRequest

The request message for EntityTypes.GetEntityType.

GetEnvironmentHistoryRequest

The request message for Environments.GetEnvironmentHistory.

GetEnvironmentRequest

The request message for Environments.GetEnvironment.

GetFulfillmentRequest

The request message for Fulfillments.GetFulfillment.

GetIntentRequest

The request message for Intents.GetIntent.

GetKnowledgeBaseRequest

Request message for KnowledgeBases.GetKnowledgeBase.

GetParticipantRequest

The request message for Participants.GetParticipant.

GetSessionEntityTypeRequest

The request message for SessionEntityTypes.GetSessionEntityType.

GetValidationResultRequest

The request message for Agents.GetValidationResult.

GetVersionRequest

The request message for Versions.GetVersion.

HumanAgentAssistantConfig

Defines the Human Agent Assist to connect to a conversation. .. attribute:: notification_config

Pub/Sub topic on which to publish new agent assistant events.

:type: google.cloud.dialogflow_v2.types.NotificationConfig

HumanAgentAssistantEvent

Represents a notification sent to Cloud Pub/Sub subscribers for human agent assistant events in a specific conversation.

HumanAgentHandoffConfig

Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access.

ImportAgentRequest

The request message for Agents.ImportAgent.

InputAudioConfig

Instructs the speech recognizer how to process the audio content.

Intent

An intent categorizes an end-user's intention for one conversation turn. For each agent, you define many intents, where your combined intents can handle a complete conversation. When an end-user writes or says something, referred to as an end-user expression or end-user input, Dialogflow matches the end-user input to the best intent in your agent. Matching an intent is also known as intent classification.

For more information, see the intent guide <https://cloud.google.com/dialogflow/docs/intents-overview>__.

IntentBatch

This message is a wrapper around a collection of intents. .. attribute:: intents

A collection of intents.

:type: Sequence[google.cloud.dialogflow_v2.types.Intent]

IntentView

Represents the options for views of an intent. An intent can be a sizable object. Therefore, we provide a resource view that does not return training phrases in the response by default.

KnowledgeBase

A knowledge base represents a collection of knowledge documents that you provide to Dialogflow. Your knowledge documents contain information that may be useful during conversations with end-users. Some Dialogflow features use knowledge bases when looking for a response to an end-user input.

For more information, see the knowledge base guide <https://cloud.google.com/dialogflow/docs/how/knowledge-bases>__.

Note: The projects.agent.knowledgeBases resource is deprecated; only use projects.knowledgeBases.

KnowledgeOperationMetadata

Metadata in google::longrunning::Operation for Knowledge operations.

ListAnswerRecordsRequest

Request message for AnswerRecords.ListAnswerRecords.

ListAnswerRecordsResponse

Response message for AnswerRecords.ListAnswerRecords.

ListContextsRequest

The request message for Contexts.ListContexts.

ListContextsResponse

The response message for Contexts.ListContexts.

ListConversationProfilesRequest

The request message for ConversationProfiles.ListConversationProfiles.

ListConversationProfilesResponse

The response message for ConversationProfiles.ListConversationProfiles.

ListConversationsRequest

The request message for Conversations.ListConversations.

ListConversationsResponse

The response message for Conversations.ListConversations.

ListDocumentsRequest

Request message for Documents.ListDocuments.

ListDocumentsResponse

Response message for Documents.ListDocuments.

ListEntityTypesRequest

The request message for EntityTypes.ListEntityTypes.

ListEntityTypesResponse

The response message for EntityTypes.ListEntityTypes.

ListEnvironmentsRequest

The request message for Environments.ListEnvironments.

ListEnvironmentsResponse

The response message for Environments.ListEnvironments.

ListIntentsRequest

The request message for Intents.ListIntents.

ListIntentsResponse

The response message for Intents.ListIntents.

ListKnowledgeBasesRequest

Request message for KnowledgeBases.ListKnowledgeBases.

ListKnowledgeBasesResponse

Response message for KnowledgeBases.ListKnowledgeBases.

ListMessagesRequest

The request message for Conversations.ListMessages.

ListMessagesResponse

The response message for Conversations.ListMessages.

ListParticipantsRequest

The request message for Participants.ListParticipants.

ListParticipantsResponse

The response message for Participants.ListParticipants.

ListSessionEntityTypesRequest

The request message for SessionEntityTypes.ListSessionEntityTypes.

ListSessionEntityTypesResponse

The response message for SessionEntityTypes.ListSessionEntityTypes.

ListVersionsRequest

The request message for Versions.ListVersions.

ListVersionsResponse

The response message for Versions.ListVersions.

LoggingConfig

Defines logging behavior for conversation lifecycle events. .. attribute:: enable_stackdriver_logging

Whether to log conversation events like CONVERSATION_STARTED to Stackdriver in the conversation project as JSON format ConversationEvent protos.

:type: bool

Message

Represents a message posted into a conversation. .. attribute:: name

The unique identifier of the message. Format: projects/<Project ID>/locations/<Location ID>/conversations/<Conversation ID>/messages/<Message ID>.

:type: str

MessageAnnotation

Represents the result of annotation for the message. .. attribute:: parts

The collection of annotated message parts ordered by their position in the message. You can recover the annotated message by concatenating [AnnotatedMessagePart.text].

:type: Sequence[google.cloud.dialogflow_v2.types.AnnotatedMessagePart]

NotificationConfig

Defines notification behavior. .. attribute:: topic

Name of the Pub/Sub topic to publish conversation events like CONVERSATION_STARTED as serialized ConversationEvent protos.

Notification works for phone calls, if this topic either is in the same project as the conversation or you grant service-<Conversation Project Number>@gcp-sa-dialogflow.iam.gserviceaccount.com the Dialogflow Service Agent role in the topic project.

Format: projects/<Project ID>/locations/<Location ID>/topics/<Topic ID>.

:type: str

OriginalDetectIntentRequest

Represents the contents of the original request that was passed to the [Streaming]DetectIntent call.

OutputAudio

Represents the natural language speech audio to be played to the end user.

OutputAudioConfig

Instructs the speech synthesizer on how to generate the output audio content. If this audio config is supplied in a request, it overrides all existing text-to-speech settings applied to the agent.

OutputAudioEncoding

Audio encoding of the output audio format in Text-To-Speech.

Participant

Represents a conversation participant (human agent, virtual agent, end-user).

QueryInput

Represents the query input. It can contain either:

  1. An audio config which instructs the speech recognizer how to process the speech audio.
  2. A conversational query in the form of text,.

  3. An event that specifies which intent to trigger.

QueryParameters

Represents the parameters of the conversational query. .. attribute:: time_zone

The time zone of this conversational query from the time zone database <https://www.iana.org/time-zones>__, e.g., America/New_York, Europe/Paris. If not provided, the time zone specified in agent settings is used.

:type: str

QueryResult

Represents the result of conversational query or event processing.

ReloadDocumentRequest

Request message for Documents.ReloadDocument.

RestoreAgentRequest

The request message for Agents.RestoreAgent.

SearchAgentsRequest

The request message for Agents.SearchAgents.

SearchAgentsResponse

The response message for Agents.SearchAgents.

Sentiment

The sentiment, such as positive/negative feeling or association, for a unit of analysis, such as the query text.

SentimentAnalysisRequestConfig

Configures the types of sentiment analysis to perform. .. attribute:: analyze_query_text_sentiment

Instructs the service to perform sentiment analysis on query_text. If not provided, sentiment analysis is not performed on query_text.

:type: bool

SentimentAnalysisResult

The result of sentiment analysis. Sentiment analysis inspects user input and identifies the prevailing subjective opinion, especially to determine a user's attitude as positive, negative, or neutral. For [Participants.DetectIntent][], it needs to be configured in DetectIntentRequest.query_params. For [Participants.StreamingDetectIntent][], it needs to be configured in StreamingDetectIntentRequest.query_params. And for Participants.AnalyzeContent and Participants.StreamingAnalyzeContent, it needs to be configured in ConversationProfile.human_agent_assistant_config

SessionEntityType

A session represents a conversation between a Dialogflow agent and an end-user. You can create special entities, called session entities, during a session. Session entities can extend or replace custom entity types and only exist during the session that they were created for. All session data, including session entities, is stored by Dialogflow for 20 minutes.

For more information, see the session entity guide <https://cloud.google.com/dialogflow/docs/entities-session>__.

SetAgentRequest

The request message for Agents.SetAgent.

SpeechContext

Hints for the speech recognizer to help with recognition in a specific conversation state.

SpeechModelVariant

Variant of the specified [Speech model][google.cloud.dialogflow.v2.InputAudioConfig.model] to use.

See the Cloud Speech documentation <https://cloud.google.com/speech-to-text/docs/enhanced-models>__ for which models have different variants. For example, the "phone_call" model has both a standard and an enhanced variant. When you use an enhanced model, you will generally receive higher quality results than for a standard model.

SpeechToTextConfig

Configures speech transcription for ConversationProfile.

SpeechWordInfo

Information for a word recognized by the speech recognizer. .. attribute:: word

The word this info is for.

:type: str

SsmlVoiceGender

Gender of the voice as described in SSML voice element <https://www.w3.org/TR/speech-synthesis11/#edef_voice>__.

StreamingDetectIntentRequest

The top-level message sent by the client to the Sessions.StreamingDetectIntent method.

Multiple request messages should be sent in order:

  1. The first message must contain session, query_input plus optionally query_params. If the client wants to receive an audio response, it should also contain output_audio_config. The message must not contain input_audio.

  2. If query_input was set to query_input.audio_config, all subsequent messages must contain input_audio to continue with Speech recognition. If you decide to rather detect an intent from text input after you already started Speech recognition, please send a message with query_input.text.

    However, note that:

    • Dialogflow will bill you for the audio duration so far.
    • Dialogflow discards all Speech recognition results in favor of the input text.
    • Dialogflow will use the language code from the first message.

After you sent all input, you must half-close or abort the request stream.

StreamingDetectIntentResponse

The top-level message returned from the StreamingDetectIntent method.

Multiple response messages can be returned in order:

  1. If the input was set to streaming audio, the first one or more messages contain recognition_result. Each recognition_result represents a more complete transcript of what the user said. The last recognition_result has is_final set to true.

  2. The next message contains response_id, query_result and optionally webhook_status if a WebHook was called.

StreamingRecognitionResult

Contains a speech recognition result corresponding to a portion of the audio that is currently being processed or an indication that this is the end of the single requested utterance.

Example:

  1. transcript: "tube"

  2. transcript: "to be a"

  3. transcript: "to be"

  4. transcript: "to be or not to be" is_final: true

  5. transcript: " that's"

  6. transcript: " that is"

  7. message_type: END_OF_SINGLE_UTTERANCE

  8. transcript: " that is the question" is_final: true

Only two of the responses contain final results (#4 and #8 indicated by is_final: true). Concatenating these generates the full transcript: "to be or not to be that is the question".

In each response we populate:

  • for TRANSCRIPT: transcript and possibly is_final.

  • for END_OF_SINGLE_UTTERANCE: only message_type.

SuggestArticlesRequest

The request message for Participants.SuggestArticles.

SuggestArticlesResponse

The response message for Participants.SuggestArticles.

SuggestFaqAnswersRequest

The request message for Participants.SuggestFaqAnswers.

SuggestFaqAnswersResponse

The request message for Participants.SuggestFaqAnswers.

SuggestionFeature

The type of Human Agent Assistant API suggestion to perform, and the maximum number of results to return for that type. Multiple Feature objects can be specified in the features list.

SuggestionResult

One response of different type of suggestion response which is used in the response of Participants.AnalyzeContent and Participants.AnalyzeContent, as well as HumanAgentAssistantEvent.

SynthesizeSpeechConfig

Configuration of how speech should be synthesized. .. attribute:: speaking_rate

Optional. Speaking rate/speed, in the range [0.25, 4.0]. 1.0 is the normal native speed supported by the specific voice. 2.0 is twice as fast, and 0.5 is half as fast. If unset(0.0), defaults to the native 1.0 speed. Any other values < 0.25 or > 4.0 will return an error.

:type: float

TextInput

Represents the natural language text to be processed. .. attribute:: text

Required. The UTF-8 encoded natural language text to be processed. Text length must not exceed 256 characters.

:type: str

TextToSpeechSettings

Instructs the speech synthesizer on how to generate the output audio content.

TrainAgentRequest

The request message for Agents.TrainAgent.

UpdateAnswerRecordRequest

Request message for AnswerRecords.UpdateAnswerRecord.

UpdateContextRequest

The request message for Contexts.UpdateContext.

UpdateConversationProfileRequest

The request message for ConversationProfiles.UpdateConversationProfile.

UpdateDocumentRequest

Request message for Documents.UpdateDocument.

UpdateEntityTypeRequest

The request message for EntityTypes.UpdateEntityType.

UpdateEnvironmentRequest

The request message for Environments.UpdateEnvironment.

UpdateFulfillmentRequest

The request message for Fulfillments.UpdateFulfillment.

UpdateIntentRequest

The request message for Intents.UpdateIntent.

UpdateKnowledgeBaseRequest

Request message for KnowledgeBases.UpdateKnowledgeBase.

UpdateParticipantRequest

The request message for Participants.UpdateParticipant.

UpdateSessionEntityTypeRequest

The request message for SessionEntityTypes.UpdateSessionEntityType.

UpdateVersionRequest

The request message for Versions.UpdateVersion.

ValidationError

Represents a single validation error. .. attribute:: severity

The severity of the error.

:type: google.cloud.dialogflow_v2.types.ValidationError.Severity

ValidationResult

Represents the output of agent validation. .. attribute:: validation_errors

Contains all validation errors.

:type: Sequence[google.cloud.dialogflow_v2.types.ValidationError]

Version

You can create multiple versions of your agent and publish them to separate environments.

When you edit an agent, you are editing the draft agent. At any point, you can save the draft agent as an agent version, which is an immutable snapshot of your agent.

When you save the draft agent, it is published to the default environment. When you create agent versions, you can publish them to custom environments. You can create a variety of custom environments for:

  • testing
  • development
  • production
  • etc.

For more information, see the versions and environments guide <https://cloud.google.com/dialogflow/docs/agents-versions>__.

VoiceSelectionParams

Description of which voice to use for speech synthesis. .. attribute:: name

Optional. The name of the voice. If not set, the service will choose a voice based on the other parameters such as language_code and ssml_gender.

:type: str

WebhookRequest

The request message for a webhook call. .. attribute:: session

The unique identifier of detectIntent request session. Can be used to identify end-user inside webhook implementation. Format: projects/<Project ID>/agent/sessions/<Session ID>, or projects/<Project ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>.

:type: str

WebhookResponse

The response message for a webhook call.

This response is validated by the Dialogflow server. If validation fails, an error will be returned in the QueryResult.diagnostic_info field. Setting JSON fields to an empty value with the wrong type is a common error. To avoid this error:

  • Use "" for empty strings
  • Use {} or null for empty objects
  • Use [] or null for empty arrays

For more information, see the Protocol Buffers Language Guide <https://developers.google.com/protocol-buffers/docs/proto3#json>__.