Human natural language is ambiguous; this forms the basis of many cultures’ sense of humor, as well as laying the foundation for interactions spanning the gamut from metaphor to double entendre. Resolving ambiguity in order to understand the intent of the speaker is a difficult task – difficult enough that human users of language sometimes fail at it, without which we would be absent many sitcom episode storylines. When dialogue systems deal with the ambiguous utterances of human users in an automated self-help environment, a model of context becomes an essential component of the understanding process.
Lexical ambiguity refers to when a given word has more than one meaning. There are hundreds of ambiguous words in English; for example, the word bank could be a place to store money, the side of a river, or what you do with a complex pool shot. In many cases, the domain of a conversation provides the necessary disambiguation:
It is highly unlikely that a user ordering pizza wants to see it decorated with an abundance of silly 1980s pop references. In well-defined task domains, disambiguation can proceed from the knowledge that the domain of the task provides a semantic neighborhood constraining the interpretation of an item. In open-ended tasks, the utterance itself provides a neighborhood; in the absence of any other context, the presence of the word pizza in that sentence can constrain the interpretation of cheese, because the non-food interpretation can be seen as too semantically distant from pizza to be plausible.
A slightly harder task is understanding the natural use of pronouns. Users of English employ the pronouns he, she, it, and they in situations when the entity or entities to which the pronoun refers should be clear. Determining what a pronoun means is called pronoun resolution. A simple example might be in the domain of reviewing a hotel reservation:
In this case, the reservation is the entity in focus; it is a trivial matter to resolve the pronoun it. Imagine, however, this interaction with a personal digital assistant:
There are two entities which the speaker could be referring to as it: the meeting and the account. However, accounts cannot be rescheduled; meetings can, and this additional semantic information allows us to decode the sentence. If the reply had been, “Who is the AE on it?” the best interpretation would be that it means the ABCD account, because accounts have account executives and meetings do not. This can be represented computationally in a semantic network that defines the relationships between concepts and their attributes. Even harder would be this interchange:
Resolving the pronoun him requires the advanced knowledge that the scheduler of the appointment would be the most appropriate person to inform about a late arrival.
Possibly the most difficult ambiguity to resolve is syntactic, which frequently involves phrases that work as modifiers and determining which part of the sentence they are modifying. For example:
Is the speaker asking the assistant to send her a reminder on Monday (when the office would be open and a call can be made), or to set up a reminder right away for altering an appointment that is occurring on Monday? That cannot be resolved with any certainty.
Resolving ambiguity typically involves the following steps: (1) Determine all possible interpretations of the sentence. (2) Leverage as much context as possible to rank these interpretations according to their likelihood. This context can be from the sentence itself, from the domain, from world knowledge, or even from user data from a resource such as CRM. (3) If there is a clearly superior choice, select that interpretation; otherwise, ask a clarifying question. While in text interfaces we want to minimize dialogue turns because they slow down the interaction, it is still important that we balance efficiency against accuracy. After all, we don’t want our interactions with dialogue systems to be fodder for any new sitcom episodes.
Latest posts by Lisa Michaud (see all)
- CXP 17: Leveraging Natural Language Understanding for Self-Service Chatbots - June 19, 2017
- Je ne parle pas le français: Why We Need Multilingual Bots - April 17, 2017
- Data Doesn’t Lie – But it Doesn’t Tell the Whole Truth - January 24, 2017