Package-level declarations
Types
Define behaviors on logging.
Hierarchical advanced settings for agent/flow/page/fulfillment/parameter. Settings exposed at lower level overrides the settings exposed at higher level. Overriding occurs at the sub-setting level. For example, the playback_interruption_settings at fulfillment level only overrides the playback_interruption_settings at the agent level, leaving other settings at the agent level unchanged. DTMF settings does not override each other. DTMF settings set at different levels define DTMF detections running in parallel. Hierarchy: Agent->Flow->Page->Fulfillment/Parameter.
Represents the natural speech audio to be processed.
One interaction between a human and virtual agent. The human provides some input and the virtual agent provides a response.
The input from the human user.
The output from the virtual agent.
Represents the input for dtmf event.
An entity entry for an associated entity type.
An excluded entity phrase that should not be matched.
The configuration for continuous tests.
Configuration for the version.
Configuration for webhooks.
An event handler specifies an event that can be handled during a session. When the specified event happens, the following actions are taken in order: * If there is a trigger_fulfillment
associated with the event, it will be called. * If there is a target_page
associated with the event, the session will transition into the specified page. * If there is a target_flow
associated with the event, the session will transition into the specified flow.
Represents the event to trigger.
Definition of the experiment.
A confidence interval is a range of possible values for the experiment objective you are trying to measure.
Metric and corresponding confidence intervals.
The inference result which includes an objective metric to optimize and the confidence interval.
Version variant and associated metrics.
Configuration for how the filling of a parameter should be handled.
Represents a form parameter.
A form is a data model that groups related parameters that can be collected from the user. The process in which the agent prompts the user and collects parameter values from the user is called form filling. A form can be added to a page. When form filling is done, the filled parameters will be written to the session.
The list of messages or conditional cases to activate for this case.
Each case has a Boolean condition. When it is evaluated to be True, the corresponding messages will be selected and evaluated recursively.
A list of cascading if-else conditions. Cases are mutually exclusive. The first one with a matching condition is selected, all the rest ignored.
A fulfillment can do one or more of the following actions at the same time: * Generate rich message responses. * Set parameter values. * Call the webhook. Fulfillments can be called at various stages in the Page or Form lifecycle. For example, when a DetectIntentRequest drives a session to enter a new page, the page's entry fulfillment can add a static response to the QueryResult in the returning DetectIntentResponse, call the webhook (for example, to load user data from a database), or both.
Setting a parameter value.
Google Cloud Storage location for a Dialogflow operation that writes or exports objects (e.g. exported agent or transcripts) outside of Dialogflow.
Instructs the speech recognizer on how to process the audio content.
Represents the intent to trigger programmatically rather than as a result of natural language processing.
Represents an intent parameter.
An intent represents a user's intent to interact with a conversational agent. You can provide information for the Dialogflow API to use to match user input to an intent by adding training phrases (i.e., examples of user input) to your intent.
Represents a part of a training phrase.
Represents an example that the agent is trained on to identify the intent.
Settings related to NLU.
A Dialogflow CX conversation (session) can be described and visualized as a state machine. The states of a CX session are represented by pages. For each flow, you define many pages, where your combined pages can handle a complete conversation on the topics the flow is designed for. At any given moment, exactly one page is the current page, the current page is considered active, and the flow associated with that page is considered active. Every flow has a special start page. When a flow initially becomes active, the start page page becomes the current page. For each conversational turn, the current page will either stay the same or transition to another page. You configure each page to collect information from the end-user that is relevant for the conversational state represented by the page. For more information, see the Page guide.
Represents the query input. It can contain one of: 1. A conversational query in the form of text. 2. An intent query that specifies which intent to trigger. 3. Natural language speech audio to be processed. 4. An event to be triggered. 5. DTMF digits to invoke an intent and fill in parameter value.
Indicates that the conversation succeeded, i.e., the bot handled the issue that the customer talked to it about. Dialogflow only uses this to determine which conversations should be counted as successful and doesn't process the metadata in this message in any way. Note that Dialogflow also considers conversations that get to the conversation end page as successful even if they don't return ConversationSuccess. You may set this, for example: * In the entry_fulfillment of a Page if entering the page indicates that the conversation succeeded. * In a webhook response when you determine that you handled the customer issue.
Indicates that interaction with the Dialogflow agent has ended. This message is generated by Dialogflow only and not supposed to be defined by the user.
Indicates that the conversation should be handed off to a live agent. Dialogflow only uses this to determine which conversations were handed off to a human agent for measurement purposes. What else to do with this signal is up to you and your handoff procedures. You may set this, for example: * In the entry_fulfillment of a Page if entering the page indicates something went extremely wrong in the conversation. * In a webhook response when you determine that the customer issue can only be handled by a human.
Represents an audio message that is composed of both segments synthesized from the Dialogflow agent prompts and ones hosted externally at the specified URIs. The external URIs are specified via play_audio. This message is generated by Dialogflow only and not supposed to be defined by the user.
Represents one segment of audio.
A text or ssml response that is preferentially used for TTS output audio synthesis, as described in the comment on the ResponseMessage message.
Specifies an audio clip to be played by the client as part of the response.
Represents a response message that can be returned by a conversational agent. Response messages are also used for output audio synthesis. The approach is as follows: * If at least one OutputAudioText response is present, then all OutputAudioText responses are linearly concatenated, and the result is used for output audio synthesis. * If the OutputAudioText responses are a mixture of text and SSML, then the concatenated result is treated as SSML; otherwise, the result is treated as either text or SSML as appropriate. The agent designer should ideally use either text or SSML consistently throughout the bot design. * Otherwise, all Text responses are linearly concatenated, and the result is used for output audio synthesis. This approach allows for more sophisticated user experience scenarios, where the text displayed to the user may differ from what is heard.
Represents the signal that telles the client to transfer the phone call connected to the agent to a third-party endpoint.
The text response message.
The configuration for auto rollout.
A single rollout step with specified traffic allocation.
State of the auto-rollout process.
Settings for exporting audio.
Settings for exporting conversations to Insights.
Settings related to speech recognition.
Represents a result from running a test case in an agent environment.
Represents configurations for a test case.
The description of differences between original and replayed agent output.
Represents the natural language text to be processed.
Settings related to speech synthesizing.
A transition route specifies a intent that can be matched and/or a data condition that can be evaluated during a session. When a specified transition is matched, the following actions are taken in order: * If there is a trigger_fulfillment
associated with the transition, it will be called. * If there is a target_page
associated with the transition, the session will transition into the specified page. * If there is a target_flow
associated with the transition, the session will transition into the specified flow.
The history of variants update.
A list of flow version variants.
A single flow version with specified traffic allocation.
Represents configuration for a generic web service.
Webhooks host the developer's business logic. During a session, webhooks allow the developer to use the data extracted by Dialogflow's natural language processing to generate dynamic responses, validate collected data, or trigger actions on the backend.
Represents configuration for a Service Directory service.
The Status
type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by gRPC. Each Status
message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the API Design Guide.