Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Dialogue system

From Wikipedia, the free encyclopedia
Computer system to converse with a human

icon
This articleneeds additional citations forverification. Please helpimprove this article byadding citations to reliable sources. Unsourced material may be challenged and removed.
Find sources: "Dialogue system" – news ·newspapers ·books ·scholar ·JSTOR
(June 2022) (Learn how and when to remove this message)
Anautomated online assistant on a website - an example where dialogue systems are major components

Adialogue system, orconversational agent (CA), is a computer system intended to converse with a human. Dialogue systems employ one or more of text, speech, graphics, haptics, gestures, and other modes for communication on both the input and output channel.

The elements of a dialogue system are not defined because this idea is under research;[citation needed] however, they are different fromchatbots.[1] The typicalGUIwizard engages in a sort of dialogue, but it includes very few of the common dialogue system components, and the dialogue state is trivial.

Background

[edit]

After dialogue systems based only on written text processing starting from the early Sixties,[2] the firstspeaking dialogue system was issued by theDARPA Project in the US in 1977.[3] After the end of this 5-year project, some European projects issued the first dialogue system able to speak many languages (also French, German and Italian).[4] Those first systems were used in the telecom industry to provide phone various services in specific domains, e.g. automated agenda and train tables service.

Components

[edit]

What sets of components are included in a dialogue system, and how those components divide up responsibilities differs from system to system. Principal to any dialogue system is thedialogue manager, which is a component that manages the state of the dialogue, and dialogue strategy. A typical activity cycle in a dialogue system contains the following phases:[5]

  1. The user speaks, and the input is converted to plain text by the system'sinput recogniser/decoder, which may include:
  2. The text is analysed by anatural language understanding (NLU) unit, which may include:
  3. The semantic information is analysed by thedialogue manager, which keeps the history and state of the dialogue and manages the general flow of the conversation.
  4. Usually, the dialogue manager contacts one or more task managers, that have knowledge of the specific task domain.
  5. The dialogue manager produces output using anoutput generator, which may include:
  6. Finally, the output is rendered using anoutput renderer, which may include:

Dialogue systems that are based on a text-only interface (e.g. text-based chat) contain only stages 2–5.

Types of systems

[edit]

Dialogue systems fall into the following categories, which are listed here along a few dimensions. Many of the categories overlap and the distinctions may not be well established.

Performance

[edit]

Some authors measure the dialogue system's performance in terms of the percentage of sentences completely right, by comparing the model of sentences (this measure is calledConcept Sentence Accuracy[6] orSentence Understanding[4]). Dialogue systems can sometimes give inconsistent responses depending on how users phrase their questions[7].

Applications

[edit]

Dialogue systems can support a broad range of applications in business enterprises, education, government, healthcare, and entertainment.[8] For example:

  • Responding to customers' questions about products and services via a company's website orintranet portal
  • Customer service agentknowledge base: Allows agents to type in a customer's question and guide them with a response
  • Guided selling: Facilitating transactions by providing answers and guidance in the sales process, particularly for complex products being sold to novice customers
  • Help desk: Responding to internal employee questions, e.g., responding to HR questions
  • Website navigation: Guiding customers to relevant portions of complex websites—a Website concierge
  • Technical support: Responding to technical problems, such as diagnosing a problem with a product or device
  • Personalized service: Conversational agents can leverage internal and external databases to personalise interactions, such as answering questions about account balances, providing portfolio information, delivering frequent flier or membership information, for example
  • Training or education: They can provide problem-solving advice while the user learns
  • Simple dialogue systems are widely used to decrease the human workload incall centers. In this and other industrial telephony applications, the functionality provided by dialogue systems is known asinteractive voice response or IVR.
  • Support scientist in data manipulation and analysis tasks, for example in genomics.[9]

In some cases, conversational agents can interact with users using artificial characters. These agents are then referred to asembodied agents.

In the 2020s, dialogue systems have increasingly been built on top of large language models (LLMs), which allow them to handle open-domain conversation more flexibly than earlier rule-based or statistical approaches.[10] Modern implementations often integrate both voice and text interfaces, providing users with multi-modal interaction through conversational agents. Such systems are also being embedded into applications with user-friendly interfaces for customer service, education, and personal assistance.[11]

Toolkits and architectures

[edit]

A survey of current frameworks, languages and technologies for defining dialogue systems.

Name & linksSystem typeDescriptionAffiliation[s]Environment[s]Comments
AIMLChatterbot languageXML dialect for creating natural language software agentsRichard Wallace, Pandorabots, Inc.
ChatScriptChatterbot languageLanguage/Engine for creating natural language software agentsBruce Wilcox
CSLU Toolkit
A state-based speech interface prototyping environmentOGI School of Science and Engineering
M. McTear
Ron Cole
publications are from 1999.
NLUI ServerDomain-independent toolkitComplete multilingual framework for buildingnatural language user interface systemsLinguaSysout-of-box support of mixed-initiative dialogues
DaVoice AIDomain-independent toolkitLanguage/Engine for creating natural language software agentsDerek Willisprimarily for telephony.
OlympusComplete framework for implementing spoken dialogue systemsCarnegie Mellon University[1]
NextnovaMultimodal PlatformPlatform for developing multimodal software applications. Based on State Chart XML (SCXML)Ponvia Technology, Inc.
VXML
Voice XML
Spoken dialogueMultimodal dialogue markup languageDeveloped initially byAT&T, then administered by an industry consortium and finally aW3C specificationExampleprimarily for telephony.
SALTmarkup languageMultimodal dialogue markup languageMicrosoft"has not reached the level of maturity of VoiceXML in the standards process".
Quack.com - QXMLDevelopment EnvironmentCompany bought byAOL
OpenDialDomain-independent toolkitHybrid symbolic/statistical framework for spoken dialogue systems, implemented in JavaUniversity of Oslo
NADIAdialogue engine and dialogue modellingCreating natural dialogues/dialogue systems. Supports dialogue acts, mixed initiative, NLG. Implemented in Java.Markus M. Bergcreate XML-based dialogue files, no need to specify grammars, publications are from 2014

See also

[edit]

References

[edit]
  1. ^Klüwer, Tina. "From chatbots to dialog systems." Conversational agents and natural language interaction: Techniques and Effective Practices. IGI Global, 2011. 1-22.
  2. ^McTear, Michael, Zoraida Callejas, and David Griol,The conversational interface: Talking to smart devices, Springer, 2016.
  3. ^Giancarlo Pirani (ed),Advanced algorithms and architectures for speech understanding, Vol. 1. Springer Science & Business Media, 2013.
  4. ^abAlberto Ciaramella,A prototype performance evaluation report, Sundial work package 8000 (1993).
  5. ^Jurafsky & Martin (2009), Speech and language processing. Pearson International Edition,ISBN 978-0-13-504196-3, Chapter 24
  6. ^Bangalore, Srinivas, and Michael Johnston. "Robust understanding in multimodal interfaces." Computational Linguistics 35.3 (2009): 345-397.
  7. ^Jurafsky, D., & Martin, J. H. *Speech and Language Processing*, 3rd edition draft. Stanford University.
  8. ^Lester, J.; Branting, K.; Mott, B. (2004),"Conversational Agents"(PDF),The Practical Handbook of Internet Computing, Chapman & Hall
  9. ^Crovari; Pidò; Pinoli; Bernasconi; Canakoglu; Garzotto; Ceri (2021), "GeCoAgent: a conversational agent for empowering genomic data extraction and analysis",ACM Transactions on Computing for Healthcare,3, ACM New York, NY:1–29,doi:10.1145/3464383,hdl:11311/1192262,S2CID 245855725
  10. ^Zhou, Kun (2024)."Large Language Models for Dialogue: A Survey".Transactions of the Association for Computational Linguistics.12:730–749.doi:10.1162/tacl_a_00686.
  11. ^"ChatGPT and the rise of AI agents".Nature. 2023-11-10. Retrieved26 August 2025.

Further reading

[edit]
Natural-language user interfaces
Graphical user interfaces
Touch user interfaces
3D user interfaces
Other user interfaces
Authority control databases: NationalEdit this at Wikidata
Retrieved from "https://en.wikipedia.org/w/index.php?title=Dialogue_system&oldid=1328341955"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp