chapter 30 advanced dialog systems - stanford …jurafsky/slp3/30.pdf · chapter 30advanced dialog...

21
Speech and Language Processing. Daniel Jurafsky & James H. Martin. Copyright c 2017. All rights reserved. Draft of August 7, 2017. CHAPTER 30 Advanced Dialog Systems A famous burlesque routine from the turn of the last century plays on the difficulty of conversational understanding by inventing a baseball team whose members have confusing names: C: I want you to tell me the names of the fellows on the St. Louis team. A: I’m telling you. Who’s on first, What’s on second, I Don’t Know is on third. C: You know the fellows’ names? A: Yes. C: Well, then, who’s playing first? A: Yes. C: I mean the fellow’s name on first. A: Who. C: The guy on first base. A: Who is on first. C: Well what are you askin’ me for? A: I’m not asking you – I’m telling you. Who is on first. Who’s on First Bud Abbott and Lou Costello’s version of an old burlesque standard. Of course outrageous names of baseball players are not a normal source of dif- ficulty in conversation. What this famous comic conversation is pointing out is that understanding and participating in dialog requires knowing whether the person you are talking to is making a statement or asking a question. Asking questions, giving orders, or making informational statements are things that people do in conversa- tion, yet dealing with these kind of actions in dialogue—what we will call dialog acts— is something that the GUS-style frame-based dialog systems of Chapter 29 are completely incapable of. In this chapter we describe the dialog-state architecture, also called the belief- state or information-state architecture. Like GUS systems, these agents fill slots, but they are also capable of understanding and generating such dialog acts, actions like asking a question, making a proposal, rejecting a suggestion, or acknowledging an utterance and they can incorporate this knowledge into a richer model of the state of the dialog at any point. Like the GUS systems, the dialog-state architecture is based on filling in the slots of frames, and so dialog-state systems have an NLU component to determine the specific slots and fillers expressed in a user’s sentence. Systems must additionally determine what dialog act the user was making, for example to track whether a user is asking a question. And the system must take into account the dialog context (what the system just said, and all the constraints the user has made in the past). Furthermore, the dialog-state architecture has a different way of deciding what to say next than the GUS systems. Simple frame-based systems often just continuously

Upload: truongnhu

Post on 23-Mar-2018

228 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

Speech and Language Processing. Daniel Jurafsky & James H. Martin. Copyright c© 2017. All

rights reserved. Draft of August 7, 2017.

CHAPTER

30 Advanced Dialog Systems

A famous burlesque routine from the turn of the last century plays on the difficultyof conversational understanding by inventing a baseball team whose members haveconfusing names:

C: I want you to tell me the names of the fellows on the St. Louis team.A: I’m telling you. Who’s on first, What’s on second, I Don’t Know is on third.C: You know the fellows’ names?A: Yes.C: Well, then, who’s playing first?A: Yes.C: I mean the fellow’s name on first.A: Who.C: The guy on first base.A: Who is on first.C: Well what are you askin’ me for?A: I’m not asking you – I’m telling you. Who is on first.

Who’s on First – Bud Abbott and Lou Costello’s version of anold burlesque standard.

Of course outrageous names of baseball players are not a normal source of dif-ficulty in conversation. What this famous comic conversation is pointing out is thatunderstanding and participating in dialog requires knowing whether the person youare talking to is making a statement or asking a question. Asking questions, givingorders, or making informational statements are things that people do in conversa-tion, yet dealing with these kind of actions in dialogue—what we will call dialogacts— is something that the GUS-style frame-based dialog systems of Chapter 29are completely incapable of.

In this chapter we describe the dialog-state architecture, also called the belief-state or information-state architecture. Like GUS systems, these agents fill slots,but they are also capable of understanding and generating such dialog acts, actionslike asking a question, making a proposal, rejecting a suggestion, or acknowledgingan utterance and they can incorporate this knowledge into a richer model of the stateof the dialog at any point.

Like the GUS systems, the dialog-state architecture is based on filling in the slotsof frames, and so dialog-state systems have an NLU component to determine thespecific slots and fillers expressed in a user’s sentence. Systems must additionallydetermine what dialog act the user was making, for example to track whether a useris asking a question. And the system must take into account the dialog context (whatthe system just said, and all the constraints the user has made in the past).

Furthermore, the dialog-state architecture has a different way of deciding what tosay next than the GUS systems. Simple frame-based systems often just continuously

Page 2: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

2 CHAPTER 30 • ADVANCED DIALOG SYSTEMS

ask questions corresponding to unfilled slots and then report back the results of somedatabase query. But in natural dialogue users sometimes take the initiative, such asasking questions of the system; alternatively, the system may not understand whatthe user said, and may need to ask clarification questions. The system needs a dialogpolicy to decide what to say (when to answer the user’s questions, when to insteadask the user a clarification question, make a suggestion, and so on).

Figure 30.1 shows a typical architecture for a dialog-state system. It has sixcomponents. As with the GUS-style frame-based systems, the speech recognitionand understanding components extract meaning from the input, and the generationand TTS components map from meaning to speech.

The parts that are different than the simple GUS system are the dialog statetracker which maintains the current state of the dialog (which include the user’smost recent dialog act, plus the entire set of slot-filler constraints the user has ex-pressed so far) and the dialog policy, which decides what the system should do orsay next.

DIALOG STATE TRACKING OVERVIEW

LEAVING FROM DOWNTOWN

LEAVING AT ONE P M

ARRIVING AT ONE P M

0.6

0.2

0.1

{ from: downtown }

{ depart-time: 1300 }

{ arrive-time: 1300 }

0.5

0.3

0.1

from: CMUto: airportdepart-time: 1300confirmed: noscore: 0.10

from: CMUto: airportdepart-time: 1300confirmed: noscore: 0.15

from: downtownto: airportdepart-time: --confirmed: noscore: 0.65

Automatic Speech Recognition (ASR)

Spoken Language Understanding (SLU)

Dialog State Tracker (DST)

Dialog Policy

act: confirmfrom: downtown

FROM DOWNTOWN, IS THAT RIGHT?

Natural Language Generation (NLG)Text to Speech (TTS)

Figure 1: Principal components of a spoken dialog system.

The topic of this paper is the dialog state tracker (DST). The DST takes as input all of the dialoghistory so far, and outputs its estimate of the current dialog state – for example, in a restaurantinformation system, the dialog state might indicate the user’s preferred price range and cuisine,what information they are seeking such as the phone number of a restaurant, and which conceptshave been stated vs. confirmed. Dialog state tracking is difficult because ASR and SLU errors arecommon, and can cause the system to misunderstand the user. At the same time, state tracking iscrucial because the dialog policy relies on the estimated dialog state to choose actions – for example,which restaurants to suggest.

In the literature, numerous methods for dialog state tracking have been proposed. These arecovered in detail in Section 3; illustrative examples include hand-crafted rules (Larsson and Traum,2000; Bohus and Rudnicky, 2003), heuristic scores (Higashinaka et al., 2003), Bayesian networks(Paek and Horvitz, 2000; Williams and Young, 2007), and discriminative models (Bohus and Rud-nicky, 2006). Techniques have been fielded which scale to realistically sized dialog problems andoperate in real time (Young et al., 2010; Thomson and Young, 2010; Williams, 2010; Mehta et al.,2010). In end-to-end dialog systems, dialog state tracking has been shown to improve overall systemperformance (Young et al., 2010; Thomson and Young, 2010).

Despite this progress, direct comparisons between methods have not been possible because paststudies use different domains and different system components for ASR, SLU, dialog policy, etc.Moreover, there has not been a standard task or methodology for evaluating dialog state tracking.Together these issues have limited progress in this research area.

The Dialog State Tracking Challenge (DSTC) series has provided a first common testbed andevaluation suite for dialog state tracking. Three instances of the DSTC have been run over a three

5

Figure 30.1 Architecture of a dialog-state system for task-oriented dialog from Williams et al. (2016).

As of the time of this writing, no commercial system uses a full dialog-state ar-chitecture, but some aspects of this architecture are beginning to appear in industrialsystems, and there are a wide variety of these systems in research labs.

Let’s turn first to a discussion of dialog acts.

30.1 Dialog Acts

A key insight into conversation—due originally to the philosopher Wittgenstein(1953) but worked out more fully by Austin (1962)—is that each utterance in adialog is a kind of action being performed by the speaker. These actions are com-monly called speech acts; here’s one taxonomy consisting of 4 major classes (Bachspeech acts

Page 3: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

30.1 • DIALOG ACTS 3

and Harnish, 1979):

Constatives: committing the speaker to something’s being the case (answering, claiming,confirming, denying, disagreeing, stating)

Directives: attempts by the speaker to get the addressee to do something (advising, ask-ing, forbidding, inviting, ordering, requesting)

Commissives: committing the speaker to some future course of action (promising, planning,vowing, betting, opposing)

Acknowledgments: express the speaker’s attitude regrading the hearer with respect to some so-cial action (apologizing, greeting, thanking, accepting an acknowledgment)

A user ordering a dialog system to do something (‘Turn up the music’) is issuinga DIRECTIVE. A user asking a question to which the system is expected to answeris also issuing a DIRECTIVE: in a sense the user is commanding the system to an-swer (‘What’s the address of the second restaurant’). By contrast, a user stating aconstraint (‘I am flying on Tuesday’) is issuing an ASSERTIVE. A user thanking thesystem is issuing an ACKNOWLEDGMENT. The dialog act expresses an importantcomponent of the intention of the speaker (or writer) in saying what they said.

While this idea of speech acts is powerful, modern systems expand these earlytaxonomies of speech acts to better describe actual conversations. This is because adialog is not a series of unrelated independent speech acts, but rather a collective actperformed by the speaker and the hearer. In performing this joint action the speakerand hearer must constantly establish common ground (Stalnaker, 1978), the set ofcommon

groundthings that are mutually believed by both speakers.

The need to achieve common ground means that the hearer must ground thegrounding

speaker’s utterances. To ground means to acknowledge, to make it clear that thehearer has understood the speaker’s meaning and intention. People need closure orgrounding for non-linguistic actions as well. For example, why does a well-designedelevator button light up when it’s pressed? Because this indicates to the elevatortraveler that she has successfully called the elevator. Clark (1996) phrases this needfor closure as follows, after Norman (1988):

Principle of closure. Agents performing an action require evidence, sufficientfor current purposes, that they have succeeded in performing it.

Grounding is also important when the hearer needs to indicate that the speakerhas not succeeded. If the hearer has problems in understanding, she must indicatethese problems to the speaker, again so that mutual understanding can eventually beachieved.

Clark and Schaefer (1989) point out a continuum of methods the hearer B canuse to ground the speaker A’s utterance, ordered from weakest to strongest:

Continued attention: B shows she is continuing to attend and therefore remains satisfied withA’s presentation.

Next contribution: B starts in on the next relevant contribution.Acknowledgment: B nods or says a continuer like uh-huh, yeah, or the like, or an assess-

ment like that’s great.Demonstration: B demonstrates all or part of what she has understood A to mean, for

example, by reformulating (paraphrasing) A’s utterance or by collabo-rative completion of A’s utterance.

Display: B displays verbatim all or part of A’s presentation.

Let’s look for examples of grounding in a conversation between a human travelagent and a human client in Fig. 30.2.

Page 4: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

4 CHAPTER 30 • ADVANCED DIALOG SYSTEMS

C1: . . . I need to travel in May.A1: And, what day in May did you want to travel?C2: OK uh I need to be there for a meeting that’s from the 12th to the 15th.A2: And you’re flying into what city?C3: Seattle.A3: And what time would you like to leave Pittsburgh?C4: Uh hmm I don’t think there’s many options for non-stop.A4: Right. There’s three non-stops today.C5: What are they?A5: The first one departs PGH at 10:00am arrives Seattle at 12:05 their time. The

second flight departs PGH at 5:55pm, arrives Seattle at 8pm. And the lastflight departs PGH at 8:15pm arrives Seattle at 10:28pm.

C6: OK I’ll take the 5ish flight on the night before on the 11th.A6: On the 11th? OK. Departing at 5:55pm arrives Seattle at 8pm, U.S. Air flight

115.C7: OK.

Figure 30.2 Part of a conversation between a travel agent (A) and client (C).

Utterance A1 shows the strongest form of grounding, in which the hearer dis-plays understanding by repeating verbatim part of the speaker’s words: in May,

This particular fragment doesn’t have an example of an acknowledgment, butthere’s an example in another fragment:

C: He wants to fly from Boston to BaltimoreA: Uh huh

The word uh-huh here is a backchannel, also called a continuer or an acknowl-backchannel

continuer edgment token. A backchannel is a (short) optional utterance that acknowledges thecontent of the utterance of the other and that doesn’t require an acknowledgment bythe other (Yngve 1970, Jefferson 1984, Schegloff 1982, Ward and Tsukahara 2000).

The third grounding method is to start in on the relevant next contribution, forexample in Fig. 30.2, where the speaker asks a question (A2) and the hearer (C3)answers it.

In a more subtle act of grounding, the speaker can combine this method with theprevious one. For example, notice that whenever the client answers a question, theagent begins the next question with And. The And indicates to the client that theagent has successfully understood the answer to the last question:

Speech acts are important for practical dialog systems, which need to distin-guish a statement from a directive, and which must distinguish (among the manykinds of directives) an order to do something from a question asking for informa-tion. Grounding is also crucial in dialog systems. Consider the unnaturalness of thisexample from Cohen et al. (2004):

(30.1) System: Did you want to review some more of your personal profile?Caller: No.System: What’s next?

Without an acknowledgment, the caller doesn’t know that the system has under-stand her ‘No’. The use of Okay below adds grounding, making (30.2) a much morenatural response than (30.1):

(30.2) System: Did you want to review some more of your personal profile?Caller: No.

Page 5: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

30.1 • DIALOG ACTS 5

System: Okay, what’s next?

Tag ExampleTHANK ThanksGREET Hello DanINTRODUCE It’s me againBYE Alright byeREQUEST-COMMENT How does that look?SUGGEST from thirteenth through seventeenth JuneREJECT No Friday I’m booked all dayACCEPT Saturday sounds fineREQUEST-SUGGEST What is a good day of the week for you?INIT I wanted to make an appointment with youGIVE REASON Because I have meetings all afternoonFEEDBACK OkayDELIBERATE Let me check my calendar hereCONFIRM Okay, that would be wonderfulCLARIFY Okay, do you mean Tuesday the 23rd?DIGRESS [we could meet for lunch] and eat lots of ice creamMOTIVATE We should go to visit our subsidiary in MunichGARBAGE Oops, I-

Figure 30.3 The 18 high-level dialog acts for a meeting scheduling task, from theVerbmobil-1 system (Jekat et al., 1995).

The ideas of speech acts and grounding are combined in a single kind of actioncalled a dialog act, a tag which represents the interactive function of the sentencedialog act

being tagged. Different types of dialog systems require labeling different kinds ofacts, and so the tagset—defining what a dialog act is exactly— tends to be designedfor particular tasks.

Figure 30.3 shows a domain-specific tagset for the task of two people schedulingmeetings. It has tags specific to the domain of scheduling, such as SUGGEST, usedfor the proposal of a particular date to meet, and ACCEPT and REJECT, used foracceptance or rejection of a proposal for a date, but also tags that have more generalfunction, like CLARIFY, used to request a user to clarify an ambiguous proposal.

Tag Sys User DescriptionHELLO(a = x,b = y, ...) X X Open a dialog and give info a = x,b = y, ...INFORM(a = x,b = y, ...) X X Give info a = x,b = y, ...REQUEST(a,b = x, ...) X X Request value for a given b = x, ...REQALTS(a = x, ...) χ X Request alternative with a = x, ...CONFIRM(a = x,b = y, ...) X X Explicitly confirm a = x,b = y, ...CONFREQ(a = x, ...,d) X χ Implicitly confirm a = x, ... and request value of dSELECT(a = x,a = y) X χ Implicitly confirm a = x, ... and request value of dAFFIRM(a = x,b = y, ...) X X Affirm and give further info a = x,b = y, ...NEGATE(a = x) χ X Negate and give corrected value a = xDENY(a = x) χ X Deny that a = xBYE() X X Close a dialog

Figure 30.4 Dialogue acts used by the HIS restaurant recommendation system of Younget al. (2010). The Sys and User columns indicate which acts are valid as system outputs anduser inputs, respectively.

Figure 30.4 shows a tagset for a restaurant recommendation system, and Fig. 30.5shows these tags labeling a sample dialog from the HIS system (Young et al., 2010).This example also shows the content of each dialog acts, which are the slot fillersbeing communicated.

Page 6: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

6 CHAPTER 30 • ADVANCED DIALOG SYSTEMS

Utterance Dialogue actU: Hi, I am looking for somewhere to eat. hello(task = find,type=restaurant)

S: You are looking for a restaurant. Whattype of food do you like?

confreq(type = restaurant, food)

U: I’d like an Italian somewhere near themuseum.

inform(food = Italian, near=museum)

S: Roma is a nice Italian restaurant nearthe museum.

inform(name = "Roma", type = restaurant,

food = Italian, near = museum)

U: Is it reasonably priced? confirm(pricerange = moderate)

S: Yes, Roma is in the moderate pricerange.

affirm(name = "Roma", pricerange =

moderate)

U: What is the phone number? request(phone)

S: The number of Roma is 385456. inform(name = "Roma", phone = "385456")

U: Ok, thank you goodbye. bye()Figure 30.5 A sample dialog from the HIS System of Young et al. (2010) using the dialog acts in Fig. 30.4.

Dialog acts don’t just appear discretely and independently; conversations havestructure, and dialogue acts reflect some of that structure. One aspect of this struc-ture comes from the field of conversational analysis or CA (Sacks et al., 1974)conversational

analysiswhich focuses on interactional properties of human conversation. CA defines ad-jacency pairs (Schegloff, 1968) as a pairing of two dialog acts, like QUESTIONSadjacency pair

and ANSWERS, PROPOSAL and ACCEPTANCE (or REJECTION), COMPLIMENTS andDOWNPLAYERS, GREETING and GREETING.

The structure, composed of a first pair part and asecond pair part, can helpdialog-state models decide what actions to take. However, dialog acts aren’t alwaysfollowed immediately by their second pair part. The two parts can be separated by aside sequence (Jefferson 1972, Schegloff 1972). One very common side sequenceside sequence

in dialog systems is the clarification question, which can form a subdialogue be-subdialogue

tween a REQUEST and a RESPONSE as in the following example caused by speechrecognition errors:

User: What do you have going to UNKNOWN WORD on the 5th?System: Let’s see, going where on the 5th?User: Going to Hong Kong.

System: OK, here are some flights...

Another kind of dialogue structure is the pre-sequence, like the following ex-pre-sequence

ample where a user starts with a question about the system’s capabilities (“Can youmake train reservations”) before making a request.

User: Can you make train reservations?System: Yes I can.User: Great, I’d like to reserve a seat on the 4pm train to New York.

A dialog-state model must be able to both recognize these kinds of structuresand make use of them in interacting with users.

Page 7: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

30.2 • DIALOG STATE: INTERPRETING DIALOGUE ACTS 7

30.2 Dialog State: Interpreting Dialogue Acts

The job of the dialog-state tracker is to determine both the current state of the frame(the fillers of each slot), as well as the user’s most recent dialog act. Note that thedialog-state includes more than just the slot-fillers expressed in the current sentence;it includes the entire state of the frame at this point, summarizing all of the user’sconstraints. The following example from Mrksic et al. (2017) shows the requiredoutput of the dialog state tracker after each turn:

User: I’m looking for a cheaper restaurantinform(price=cheap)

System: Sure. What kind - and where?User: Thai food, somewhere downtown

inform(price=cheap, food=Thai, area=centre)

System: The House serves cheap Thai foodUser: Where is it?

inform(price=cheap, food=Thai, area=centre); request(address)

System: The House is at 106 Regent Street

How can we interpret a dialog act, deciding whether a given input is a QUES-TION, a STATEMENT, or a SUGGEST (directive)? Surface syntax seems like a use-ful cue, since yes-no questions in English have aux-inversion (the auxiliary verbprecedes the subject), statements have declarative syntax (no aux-inversion), andcommands have no syntactic subject:

(30.3) YES-NO QUESTION Will breakfast be served on USAir 1557?STATEMENT I don’t care about lunch.COMMAND Show me flights from Milwaukee to Orlando.

Alas, the mapping from surface form to dialog act is complex. For example, thefollowing utterance looks grammatically like a YES-NO QUESTION meaning some-thing like Are you capable of giving me a list of. . . ?:

(30.4) Can you give me a list of the flights from Atlanta to Boston?

In fact, however, this person was not interested in whether the system was capa-ble of giving a list; this utterance was a polite form of a REQUEST, meaning some-thing like Please give me a list of. . . . What looks on the surface like a QUESTIONcan really be a REQUEST.

Conversely, what looks on the surface like a STATEMENT can really be a QUES-TION. The very common CHECK question (Carletta et al. 1997, Labov and Fan-shel 1977) asks an interlocutor to confirm something that she has privileged knowl-edge about. CHECKS have declarative surface form:

A OPEN-OPTION I was wanting to make some arrangements for a trip that I’m goingto be taking uh to LA uh beginning of the week after next.

B HOLD OK uh let me pull up your profile and I’ll be right with you here.[pause]

B CHECK And you said you wanted to travel next week?A ACCEPT Uh yes.

Utterances that use a surface statement to ask a question or a surface questionto issue a request are called indirect speech acts. These indirect speech acts have aindirect speech

act

Page 8: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

8 CHAPTER 30 • ADVANCED DIALOG SYSTEMS

rich literature in philosophy, but viewed from the perspective of dialog understand-ing, indirect speech acts are merely one instance of the more general problem ofdetermining the dialog act function of a sentence.

Many features can help in this task. To give just one example, in spoken-language systems, prosody or intonation (Chapter ??) is a helpful cue. Prosodyprosody

intonation or intonation is the name for a particular set of phonological aspects of the speechsignal the tune and other changes in the pitch (which can be extracted from the fun-damental frequency F0) the accent, stress, or loudness (which can be extracted fromenergy), and the changes in duration and rate of speech. So, for example, a risein pitch at the end of the utterance is a good cue for a YES-NO QUESTION, whiledeclarative utterances (like STATEMENTS) have final lowering: a drop in F0 at thefinal lowering

end of the utterance.

30.2.1 Sketching an algorithm for dialog act interpretationSince dialog acts places some constraints on the slots and values, the tasks of dialog-act detection and slot-filling are often performed jointly. Consider the task of deter-mining that

I’d like Cantonese food near the Mission District

has the structure

inform(food=cantonese,area=mission)).

The joint dialog act interpretation/slot filling algorithm generally begins witha first pass classifier to decide on the dialog act for the sentence. In the case ofthe example above, this classifier would choosing inform from among the set ofpossible dialog acts in the tag set for this particular task. Dialog act interpretation isgenerally modeled as a supervised classification task, trained on a corpus in whicheach utterance is hand-labeled for its dialog act, and relying on a wide variety offeatures, including unigrams and bigrams (show me is a good cue for a REQUEST, arethere for a QUESTION), parse features, punctuation, dialog context, and the prosodicfeatures described above.

A second pass classifier might use any of the algorithms for slot-filler extractiondiscussed in Section ?? of Chapter 29, such as CRF or RNN-based IOB tagging.Alternatively, a multinominal classifier can be used to choose between all possibleslot-value pairs, again using any of the feature functions defined in Chapter 29. Thisis possible since the domain ontology for the system is fixed, so there is a finitenumber of slot-value pairs.

Both classifiers can be built from any standard multinominal classifier (logis-tic regression, SVM), using the various features described above, or, if sufficienttraining data is available, can be built with end-to-end neural models.

30.2.2 A special case: detecting correction actsSome dialog acts are important because of their implications for dialog control. If adialog system misrecognizes or misunderstands an utterance, the user will generallycorrect the error by repeating or reformulating the utterance. Detecting these usercorrection acts is therefore quite important. Ironically, it turns out that correctionsuser correction

actsare actually harder to recognize than normal sentences! In fat, corrections in oneearly dialog system (the TOOT system) had double the ASR word error rate of non-corrections Swerts et al. (2000)! One reason for this is that speakers sometimesuse a specific prosodic style for corrections called hyperarticulation, in which thehyperarticula-

tion

Page 9: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

30.3 • DIALOGUE POLICY 9

utterance contains some exaggerated energy, duration, or F0 contours, such as I saidBAL-TI-MORE, not Boston (Wade et al. 1992, Levow 1998, Hirschberg et al. 2001).Even when they are not hyperarticulating, users who are frustrated seem to speak ina way that is harder for speech recognizers (Goldberg et al., 2003).

What are the characteristics of these corrections? User corrections tend to beeither exact repetitions or repetitions with one or more words omitted, although theymay also be paraphrases of the original utterance. (Swerts et al., 2000). Detectingthese reformulations or correction acts can be done by any classifier; some stan-dard features used for this task are shown below (Levow 1998, Litman et al. 1999,Hirschberg et al. 2001, Bulyko et al. 2005, Awadallah et al. 2015):

lexical features words like “no”, “correction”, “I don’t”, or even swear words, utterance lengthsemantic features overlap between the candidate correction act and the user’s prior utterance (computed

by word overlap or via cosines over embedding vectors)phonetic features phonetic overlap between the candidate correction act and the user’s prior utterance

(i.e. “WhatsApp” may be incorrectly recognized as “What’s up”)prosodic features hyperarticulation, increases in F0 range, pause duration, and word duration, generally

normalized by the values for previous sentencesASR features ASR confidence, language model probability

30.3 Dialogue Policy

The goal of the dialog policy is to decide what action the system should take next,dialog policy

that is, what dialog act to generate. We begin in the next section by introducing onespecific dialog policy decision, relating to confirmation: how we confirm to the userwhat we think she said. We then sketch a basic policy algorithm that could applyto all decisions. Finally, once a speech act has been generated, the natural languagegeneration component needs to generate the text of a response to the user.

30.3.1 Generating Dialogue Acts: Confirmation and RejectionModern dialog systems often make mistakes. It is therefore important for dialogsystems to make sure that they have achieved the correct interpretation of the user’sinput. This is generally done by two methods: confirming understandings with theuser and rejecting utterances that the system is likely to have misunderstood.

Various strategies can be employed for confirmation with the user. When usingthe explicit confirmation strategy, a system asks the user a direct question to con-explicit

confirmationfirm the system’s understanding, like the two examples below in which the systemasks a (boldface) yes-no confirmation questions:

S: Which city do you want to leave from?U: Baltimore.S: Do you want to leave from Baltimore?U: Yes.U: I’d like to fly from Denver Colorado to New York City on September

twenty first in the morning on United AirlinesS: Let’s see then. I have you going from Denver Colorado to New York

on September twenty first. Is that correct?U: Yes

Page 10: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

10 CHAPTER 30 • ADVANCED DIALOG SYSTEMS

When using the implicit confirmation strategy, a system instead uses the demon-implicitconfirmation

stration or display grounding strategies described above, repeating back the system’sunderstanding as part of asking the next question, as in the two examples below:

U: I want to travel to BerlinS: When do you want to travel to Berlin?U2: Hi I’d like to fly to Seattle Tuesday MorningA3: Traveling to Seattle on Tuesday, August eleventh in the morning.

Your full name?

Explicit and implicit confirmation have complementary strengths. Explicit con-firmation makes it easier for users to correct the system’s misrecognitions since auser can just answer “no” to the confirmation question. But explicit confirmation isawkward and increases the length of the conversation (Danieli and Gerbino 1995,Walker et al. 1998). The explicit confirmation dialog fragments above sound non-natural and definitely non-human; implicit confirmation is much more conversation-ally natural.

Confirmation is just one kind of conversational action by which a system canexpress lack of understanding. Another option is rejection, in which a system givesrejection

the user a prompt like I’m sorry, I didn’t understand that.Sometimes utterances are rejected multiple times. This might mean that the user

is using language that the system is unable to follow. Thus, when an utterance isrejected, systems often follow a strategy of progressive prompting or escalatingprogressive

promptingdetail (Yankelovich et al. 1995, Weinschenk and Barker 2000), as in this examplefrom Cohen et al. (2004):

System: When would you like to leave?Caller: Well, um, I need to be in New York in time for the first World Series game.System: <reject>. Sorry, I didn’t get that. Please say the month and day you’d like

to leave.Caller: I wanna go on October fifteenth.

In this example, instead of just repeating “When would you like to leave?”, therejection prompt gives the caller more guidance about how to formulate an utter-ance the system will understand. These you-can-say help messages are important inhelping improve systems’ understanding performance (Bohus and Rudnicky, 2005).If the caller’s utterance gets rejected yet again, the prompt can reflect this (“I stilldidn’t get that”), and give the caller even more guidance.

An alternative strategy for error handling is rapid reprompting, in which therapidreprompting

system rejects an utterance just by saying “I’m sorry?” or “What was that?” Onlyif the caller’s utterance is rejected a second time does the system start applyingprogressive prompting. Cohen et al. (2004) summarize experiments showing thatusers greatly prefer rapid reprompting as a first-level error prompt.

Various factors can be used as features to the dialog policy in deciding whetherto use explicit confirmation, implicit confirmation, or rejection. For example, theconfidence that the ASR system assigns to an utterance can be used by explicitlyconfirming low-confidence sentences. Recall from page ?? that confidence is a met-ric that the speech recognizer can assign to its transcription of a sentence to indi-cate how confident it is in that transcription. Confidence is often computed fromthe acoustic log-likelihood of the utterance (greater probability means higher confi-dence), but prosodic features can also be used in confidence prediction. For example,

Page 11: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

30.4 • A SIMPLE POLICY BASED ON LOCAL CONTEXT 11

utterances with large F0 excursions or longer durations, or those preceded by longerpauses, are likely to be misrecognized (Litman et al., 2000).

Another common feature in confirmation is the cost of making an error. For ex-ample, explicit confirmation is common before a flight is actually booked or moneyin an account is moved. Systems might have a four-tiered level of confidence withthree thresholds α , β , and γ:

< α low confidence reject≥ α above the threshold confirm explicitly≥ β high confidence confirm implictly≥ γ very high confidence don’t confirm at all

30.4 A simple policy based on local context

The goal of the dialog policy at turn i in the conversation is to predict which actionAi to take, based on the entire dialog state. The state could mean the entire sequenceof dialog acts from the system (A) and from the user (U), in which case the taskwould be to compute:

Ai = argmaxAi∈A

P(Ai|(A1,U1, ...,Ai−1,Ui−1) (30.5)

We can simplify this by maintaining as the dialog state mainly just the set ofslot-fillers that the user has expressed, collapsing across the many different conver-sational paths that could lead to the same set of filled slots.

Such a policy might then just condition on the current state of the frame Framei(which slots are filled and with what) and the last turn by the system and user:

Ai = argmaxAi∈A

P(Ai|Framei−1,Ai−1,Ui−1) (30.6)

Given a large enough corpus of conversations, these probabilities can be esti-mated by a classifier. Getting such enormous amounts of data can be difficult, andoften involves building user simulators to generate artificial conversations to trainon.

30.5 Natural language generation in the dialog-state model

Once a dialog act has been decided, we need to generate the text of the responseto the user. The task of natural language generation (NLG) in the information-statearchitecture is often modeled in two stages, content planning (what to say), andcontent

planningsentence realization (how to say it).sentence

realizationHere we’ll assume content planning has been done by the dialog policy, which

has chosen the dialog act to generate, and perhaps also chosen some some additionalattributes (slots and values) that the planner wants to implicitly confirm to the user.Fig. 30.6 shows a sample input structure from the policy/content planner, and oneexample of a resulting sentence that the sentence realizer could generate from thisstructure.

Let’s walk through the sentence realization stage for the example in Fig. 30.6,which comes from the classic information state statistical NLG system of Oh and

Page 12: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

12 CHAPTER 30 • ADVANCED DIALOG SYSTEMS

4.1. Template-based generation

Our NLG module started off with around 50 templates. The number of templatesgrew as we added more functionality to our system. The largest expansion camewith the addition of a ‘‘help’’ speech act, which added 16 templates to provide context-sensitive help messages. Additional information about the template system is available inOh (2000). Note that templates are not simple sentence frames with variable slots. Theyalso need to include a computational component that deals with options For example, forthe template ‘‘What time would you like to travel from {departure_city} on {depar-ture_date}?’’, if the input frame did not contain values for the attributes {departure_city}and {departure_date}, instead of generating the ungrammatical sentence ‘‘What timewould you like to travel from on?’’, it would generate ‘‘What time would you like totravel?’’. This reduces the number of templates significantly, but only at the expense ofintroducing more complexity to the templates, especially for templates that can have asmany as ten different attributes. Hence, the amount of time the developer needs to spendon crafting and maintaining the templates does not decrease significantly. At one point,the number of templates grew to nearly one hundred, some of them very complex andcumbersome to maintain. Axelrod (2000) has alluded to similar requirements in thesystem that he has described.

4.2. Development of corpus-based stochastic generator

What is perhaps more important than reducing development time is being able togenerate utterances that promote a natural interaction with the user. One of the diffi-culties for a template writer is choosing the right words, the template system’s equiv-alent of lexical selection. Often, the words that the template writer chooses for a givenutterance are different from what the domain expert would use. This mismatch mayhamper a smooth interaction because when a system utterance contains unfamiliarwords in that domain, not only does it sound unnatural, but it may also lead the user toconfusion or an inappropriate response.

One solution might be to base the generator on a corpus of task-oriented human–human conversations between a domain expert and a client. We could, for example,take the expert’s utterances and use them directly as templates. This is very simple, butis not practical, as one would need to find an utterance for every possible combinationof attributes.

The statistical n-gram language model provides an alternative representation. Then-gram language model has the advantage that it is simple to build and understand, and

Figure 3. An input frame to NLG in the Communicator.

392 A.H. Oh and A.I. Rudnicky

Figure 30.6 An input frame to NLG and a resulting output sentence, in the Communicatorsystem of Oh and Rudnicky (2000).

query arrive city hotel hotel chain inform flight earlierquery arrive time hotel hotel info inform flight earliestquery confirm hotel need car inform flight laterquery depart date hotel need hotel inform flight latestquery depart time hotel where inform flight returningquery pay by card inform airport inform not availquery preferred airport inform confirm utterance inform num flightsquery return date inform epilogue inform pricequery return time inform flight otherhotel car info inform flight another

Figure 30.7 Dialog acts in the CMU communicator system of Oh and Rudnicky (2000).

Rudnicky (2000), part of the CMU Communicator travel planning dialog system.Notice first that the policy has decided to generate the dialog act QUERY with theargument DEPART TIME. Fig. 30.7 lists the dialog acts in the Oh and Rudnicky(2000) system, each of which combines an act with a potential argument. The inputframe in Fig. 30.6 also specifies some additional filled slots that should be includedin the sentence to the user (depart airport BOS, and the depart date).

The sentence realizer acts in two steps. It will first generate a delexicalizeddelexicalized

string like:

What time on [depart date] would you like to leave [depart airport]?

Delexicalization is the process of replacing specific words with a generic rep-resentation of their slot types. A delexicalized sentence is much easier to generatesince we can train on many different source sentences from different specific datesand airports. Then once we’ve generating the delexicalized string, we can simply usethe input frame from the content planner to relexicalize (fill in the exact departurerelexicalize

date and airport).To generate the delexicalized sentences, the sentence realizer uses a large corpus

of human-human travel dialogs that were labeled with the dialog acts from Fig. 30.7and the slots expressed in each turn, like the following:

QUERY DEPART TIME And what time would you like to leave [depart city Pittsburgh]?QUERY ARRIVE CITY And you’re flying into what city?QUERY ARRIVE TIME What time on [arrive date May 5]?INFORM FLIGHT The flight departs [depart airport PGH] at [depart time 10 am] and arrives

[arrive city Seattle] at [arrive time 12:05 their time].

This corpus is then delexicalized, and divided up into separate corpora for eachdialog act. Thus the delexicalized corpus for one dialog act, QUERY DEPART TIMEmight be trained on examples like:

Page 13: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

30.5 • NATURAL LANGUAGE GENERATION IN THE DIALOG-STATE MODEL 13

And what time would you like to leave depart city?When would you like to leave depart city?When would you like to leave?What time do you want to leave on depart date?OK, on depart date, what time do you want to leave?

A distinct N-gram grammar is then trained for each dialog act. Now, giventhe dialog act QUERY DEPART TIME, the system samples random sentences fromthis language model. Recall from the the ”Shannon” exercise of ?? that this works(assuming a bigram LM) by first selecting a bigram (< s >,< w >) according to itsbigram probability in the language model, then drawing a bigram starting with <w > according to its bigram probability, and so on until a full sentence is generated.The probability of each successive word wi being generated from utterance class uis thus

P(wi) = P(wi|wi−1,wi−2, ...,wi−(n−1),u) (30.7)

Each of these randomly sampled sentences is then assigned a score based onheuristic rules that penalize sentences that are too short or too long, repeat slots, orlack some of the required slots from the input frame (in this case, depart airport anddepart date). The best scoring sentence is then chosen. Let’s suppose in this casewe produce the following (delexicalized) sentence:

What time on depart date would you like to leave depart airport?

This sentence is then relexicalized from the true values in the input frame, re-sulting in the final sentence:

What time on October fifth would you like to leave Boston?

More recent work has replaced the simplistic N-gram part of the generator withneural models, which similarly learn to map from an input frame to a resulting sen-tence (Wen et al. 2015a, Wen et al. 2015b).

It’s also possible to design NLG algorithms that are specific to a particular di-alog act. For example, consider the task of generating clarification questions, inclarification

questionscases where the speech recognition fails to understand some part of the user’s ut-terance. While it is possible to use the generic dialog act REJECT (“Please repeat”,or “I don’t understand what you said”), studies of human conversations show thathumans instead use targeted clarification questions that reprise elements of the mis-understanding (Purver 2004, Ginzburg and Sag 2000, Stoyanchev et al. 2013).

For example, in the following hypothetical example the system reprises thewords “going” and “on the 5th” to make it clear which aspect of the user’s turnthe system needs to be clarified:

User: What do you have going to UNKNOWN WORD on the 5th?System: Going where on the 5th?

Targeted clarification questions can be created by rules (such as replacing “go-ing to UNKNOWN WORD” with “going where”) or by building classifiers to guesswhich slots might have been misrecognized in the sentence (Chu-Carroll and Car-penter 1999, Stoyanchev et al. 2014, Stoyanchev and Johnston 2015).

Page 14: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

14 CHAPTER 30 • ADVANCED DIALOG SYSTEMS

30.6 Advanced: Markov Decision Processes

The policy we described in Section 30.4, deciding what actions the system shouldtake based just on the current filled slots and the users last utterance, has a problem:it looks only at the past of the dialog, completely ignoring whether the action wetake is likely to lead to a successful outcome (a correctly booked flight or filled-incalendar).

But we can’t know whether the outcome is successful until long after the currentutterance we are trying to plan. Reinforcement learning is the branch of machinelearning that deals with models that learn to maximize future rewards.

This is an extremely active area of research, so we give here just the simplestintuition for this direction, based on an oversimplified model of dialog as a Markovdecision process.Markov

decision processA Markov decision process or MDP is characterized by a set of states S an agentMDP

can be in, a set of actions A the agent can take, and a reward r(a,s) that the agentreceives for taking an action in a state. Given these factors, we can compute a policyπ that specifies which action a the agent should take when in a given state s so as toreceive the best reward.

To understand each of these components, we need to look at a tutorial examplein which the state space is extremely reduced. Let’s look at a trivial pedagogicalframe-and-slot example from Levin et al. (2000), a “Day-and-Month” dialog systemwhose goal is to get correct values of day and month for a two-slot frame throughthe shortest possible interaction with the user.

In principle, a state of an MDP could include any possible information about thedialog, such as the complete dialog history so far. Using such a rich model of statewould make the number of possible states extraordinarily large. So a model of stateis usually chosen that encodes a much more limited set of information, such as thevalues of the slots in the current frame, the most recent question asked to the user, theuser’s most recent answer, the ASR confidence, and so on. For the Day-and-Monthexample, let’s represent the state of the system as the values of the two slots day andmonth. There are 411 states (366 states with a day and month (counting leap year),12 states with a month but no day (d = 0, m = 1,2, ...,12), 31 states with a day butno month (m = 0, d = 1,2, ...,31), and a special initial state si and final state s f .

Actions of an MDP dialog system might include generating particular speechacts, or performing a database query to find out information. For the Day-and-Monthexample, Levin et al. (2000) propose the following actions:

• ad : a question asking for the day• am: a question asking for the month• adm: a question asking for both the day and the month• a f : a final action submitting the form and terminating the dialog

Since the goal of the system is to get the correct answer with the shortest inter-action, one possible reward function for the system would integrate three terms:

R =−(wini +wene +w f n f ) (30.8)

The term ni is the number of interactions with the user, ne is the number of errors,n f is the number of slots that are filled (0, 1, or 2), and the ws are weights.

Finally, a dialog policy π specifies which actions to apply in which state. Con-sider two possible policies: (1) asking for day and month separately, and (2) askingfor them together. These might generate the two dialogs shown in Fig. 30.8.

Page 15: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

30.6 • ADVANCED: MARKOV DECISION PROCESSES 15

d=0m=0

d=Dm=0

d=Dm=M

d=-1m=-1

d=Dm=M

d=-1m=-1

d=0m=0

Which day? Which month?

What date? Goodbye.

Goodbye.

Policy 1 (directive)

Policy 2 (open)

c1 = -3wi + 2pdwe

c2 = -2wi + 2powe

Figure 30.8 Two policies for getting a month and a day. After Levin et al. (2000).

In policy 1, the action specified for the no-date/no-month state is to ask for a day,and the action specified for any of the 31 states where we have a day but not a monthis to ask for a month. In policy 2, the action specified for the no-date/no-month stateis to ask an open-ended question (Which date) to get both a day and a month. Thetwo policies have different advantages; an open prompt can lead to shorter dialogsbut is likely to cause more errors, whereas a directive prompt is slower but lesserror-prone. Thus, the optimal policy depends on the values of the weights w andalso on the error rates of the ASR component. Let’s call pd the probability of therecognizer making an error interpreting a month or a day value after a directiveprompt. The (presumably higher) probability of error interpreting a month or dayvalue after an open prompt we’ll call po. The reward for the first dialog in Fig. 30.8is thus −3×wi + 2× pd ×we. The reward for the second dialog in Fig. 30.8 is−2×wi + 2× po×we. The directive prompt policy, policy 1, is thus better thanpolicy 2 when the improved error rate justifies the longer interaction, that is, whenpd− po >

wi2we

.In the example we’ve seen so far, there were only two possible actions, and hence

only a tiny number of possible policies. In general, the number of possible actions,states, and policies is quite large, and so the problem of finding the optimal policyπ∗ is much harder.

Markov decision theory together with classical reinforcement learning gives usa way to think about this problem. First, generalizing from Fig. 30.8, we can thinkof any particular dialog as a trajectory in state space:

s1→a1,r1 s2→a2,r2 s3→a3,r3 · · · (30.9)

The best policy π∗ is the one with the greatest expected reward over all trajec-tories. What is the expected reward for a given state sequence? The most commonway to assign utilities or rewards to sequences is to use discounted rewards. Herediscounted

rewardwe compute the expected cumulative reward Q of a sequence as a discounted sumof the utilities of the individual states:

Q([s0,a0,s1,a1,s2,a2 · · ·]) = R(s0,a0)+ γR(s1,a1)+ γ2R(s2,a2)+ · · · , (30.10)

The discount factor γ is a number between 0 and 1. This makes the agent caremore about current rewards than future rewards; the more future a reward, the morediscounted its value.

Given this model, it is possible to show that the expected cumulative rewardQ(s,a) for taking a particular action from a particular state is the following recursiveequation called the Bellman equation:Bellman

equation

Page 16: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

16 CHAPTER 30 • ADVANCED DIALOG SYSTEMS

Q(s,a) = R(s,a)+ γ

∑s′

P(s′|s,a)maxa′

Q(s′,a′) (30.11)

What the Bellman equation says is that the expected cumulative reward for agiven state/action pair is the immediate reward for the current state plus the expecteddiscounted utility of all possible next states s′, weighted by the probability of movingto that state s′, and assuming that once there we take the optimal action a′.

Equation 30.11 makes use of two parameters. We need a model of P(s′|s,a),that is, how likely a given state/action pair (s,a) is to lead to a new state s′. Andwe also need a good estimate of R(s,a). If we had lots of labeled training data, wecould simply compute both of these from labeled counts. For example, with labeleddialogs, to estimate P(s′|s,a) we could simply count how many times we were ina given state s, and out of that how many times we took action a to get to state s′.Similarly, if we had a hand-labeled reward for each dialog, we could build a modelof R(s,a).

Given these parameters, there is an iterative algorithm for solving the Bellmanequation and determining proper Q values, the value iteration algorithm (Suttonvalue iteration

and Barto 1998, Bellman 1957). See Russell and Norvig (2002) for the details ofthe algorithm.

How do we get enough labeled training data to set these parameters? This isespecially worrisome since in real problems the number of states s is extremelylarge. The most common method is to build a simulated user. The user interacts withthe system millions of times, and the system learns the state transition and rewardprobabilities from this corpus. For example Levin et al. (2000) build a generativestochastic model that given the system’s current state and actions, produced a frame-slot representation of a user response; the parameters of the simulated user wereestimated from a corpus of ATIS dialogs.

The MDP is only useful in small toy examples and is not used in practical dialogsystems. A more powerful model, the partially observable Markov decision process,or POMDP, adds extra latent variables to represent our uncertainty about the truestate of the dialog. Both MDPs and POMDPs, however, have problems due to com-putational complexity and due to their reliance on simulations that don’t reflect trueuser behavior.

Recent research has therefore focused on ways to build real task-based systemsthat nonetheless make use of this reinforcement learning intuition, often by addingreinforcement learning to deep neural networks. This is an exciting new area ofresearch, but a standard paradigm has yet to emerge.

30.7 Summary

• In dialog, speaking is a kind of action; these acts are referred to as speechacts. Speakers also attempt to achieve common ground by acknowledgingthat they have understand each other. The dialog act combines the intuitionof speech acts and grounding acts.

• The dialog-state or information-state architecture augments the frame-and-slot state architecture by keeping track of user’s dialog acts and includes apolicy for generating its own dialog acts in return.

• Policies based on reinforcement learning architecture like the MDP and POMDP

Page 17: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

BIBLIOGRAPHICAL AND HISTORICAL NOTES 17

offer ways for future dialog reward to be propagated back to influence policyearlier in the dialog manager.

Bibliographical and Historical NotesThe idea that utterances in a conversation are a kind of action being performed bythe speaker was due originally to the philosopher Wittgenstein (1953) but worked outmore fully by Austin (1962) and his student John Searle. Various sets of speech actshave been defined over the years, and a rich linguistic and philosophical literaturedeveloped, especially focused on explaining the use of indirect speech acts.

The idea of dialog acts draws also from a number of other sources, includingthe ideas of adjacency pairs, pre-sequences, and other aspects of the internationalproperties of human conversation developed in the field of conversation analysis(see Levinson (1983) for an introduction to the field).

This idea that acts set up strong local dialogue expectations was also prefiguredby Firth (1935, p. 70), in a famous quotation:

Most of the give-and-take of conversation in our everyday life is stereotypedand very narrowly conditioned by our particular type of culture. It is a sortof roughly prescribed social ritual, in which you generally say what the otherfellow expects you, one way or the other, to say.

Another important research thread modeled dialog as a kind of collaborative be-havior, including the ideas of common ground (Clark and Marshall, 1981), referenceas a collaborative process (Clark and Wilkes-Gibbs, 1986), joint intention (Levesqueet al., 1990), and shared plans (Grosz and Sidner, 1980).

The information state model of dialogue was also strongly informed by analyticwork on the linguistic properties of dialog acts and on methods for their detection(Sag and Liberman 1975, Hinkelman and Allen 1989, Nagata and Morimoto 1994,Goodwin 1996, Chu-Carroll 1998, Shriberg et al. 1998, Stolcke et al. 2000, Gravanoet al. 2012).

Two important lines of research focused on the computational properties of con-versational structure. One line, first suggested at by Bruce (1975), suggested thatsince speech acts are actions, they should be planned like other actions, and drewon the AI planning literature (Fikes and Nilsson, 1971). An agent seeking to findout some information can come up with the plan of asking the interlocutor for theinformation. An agent hearing an utterance can interpret a speech act by running theplanner “in reverse”, using inference rules to infer from what the interlocutor saidwhat the plan might have been. Plan-based models of dialogue are referred to as BDIBDI

models because such planners model the beliefs, desires, and intentions (BDI) ofthe agent and interlocutor. BDI models of dialogue were first introduced by Allen,Cohen, Perrault, and their colleagues in a number of influential papers showing howspeech acts could be generated (Cohen and Perrault, 1979) and interpreted (Perraultand Allen 1980, Allen and Perrault 1980). At the same time, Wilensky (1983) intro-duced plan-based models of understanding as part of the task of interpreting stories.

Another influential line of research focused on modeling the hierarchical struc-ture of dialog. Grosz’s pioneering (1977) dissertation first showed that “task-orienteddialogs have a structure that closely parallels the structure of the task being per-formed” (p. 27), leading to her work with Sidner and others showing how to use

Page 18: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

18 CHAPTER 30 • ADVANCED DIALOG SYSTEMS

similar notions of intention and plans to model discourse structure and coherence indialogue. See, e.g., Lochbaum et al. (2000) for a summary of the role of intentionalstructure in dialog.

The idea of applying reinforcement learning to dialogue first came out of AT&Tand Bell Laboratories around the turn of the century with work on MDP dialoguesystems (Walker 2000, Levin et al. 2000, Singh et al. 2002) and work on cue phrases,prosody, and rejection and confirmation. Reinforcement learning research turnedquickly to the more sophisticated POMDP models (Roy et al. 2000, Lemon et al. 2006,Williams and Young 2007) applied to small slot-filling dialogue tasks.

More recent work has applied deep learning to many components of dialoguesystems.

Page 19: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

Bibliographical and Historical Notes 19

Allen, J. and Perrault, C. R. (1980). Analyzing intention inutterances. Artificial Intelligence, 15, 143–178.

Austin, J. L. (1962). How to Do Things with Words. HarvardUniversity Press.

Awadallah, A. H., Kulkarni, R. G., Ozertem, U., and Jones,R. (2015). Charaterizing and predicting voice query refor-mulation. In CIKM-15.

Bach, K. and Harnish, R. (1979). Linguistic communicationand speech acts. MIT Press.

Bellman, R. (1957). Dynamic Programming. Princeton Uni-versity Press.

Bohus, D. and Rudnicky, A. I. (2005). Sorry, I didn’t catchthat! — An investigation of non-understanding errors andrecovery strategies. In Proceedings of SIGDIAL, Lisbon,Portugal.

Bruce, B. C. (1975). Generation as a social action. In Pro-ceedings of TINLAP-1 (Theoretical Issues in Natural Lan-guage Processing), pp. 64–67. Association for Computa-tional Linguistics.

Bulyko, I., Kirchhoff, K., Ostendorf, M., and Goldberg, J.(2005). Error-sensitive response generation in a spokenlanguage dialogue system. Speech Communication, 45(3),271–288.

Carletta, J., Isard, A., Isard, S., Kowtko, J. C., Doherty-Sneddon, G., and Anderson, A. H. (1997). The reliabil-ity of a dialogue structure coding scheme. ComputationalLinguistics, 23(1), 13–32.

Chu-Carroll, J. (1998). A statistical model for discourse actrecognition in dialogue interactions. In Chu-Carroll, J. andGreen, N. (Eds.), Applying Machine Learning to DiscourseProcessing. Papers from the 1998 AAAI Spring Symposium.Tech. rep. SS-98-01, pp. 12–17. AAAI Press.

Chu-Carroll, J. and Carpenter, B. (1999). Vector-based nat-ural language call routing. Computational Linguistics,25(3), 361–388.

Clark, H. H. (1996). Using Language. Cambridge UniversityPress.

Clark, H. H. and Marshall, C. (1981). Definite reference andmutual knowledge. In Joshi, A. K., Webber, B. L., and Sag,I. A. (Eds.), Elements of Discourse Understanding, pp. 10–63. Cambridge.

Clark, H. H. and Schaefer, E. F. (1989). Contributing to dis-course. Cognitive Science, 13, 259–294.

Clark, H. H. and Wilkes-Gibbs, D. (1986). Referring as acollaborative process. Cognition, 22, 1–39.

Cohen, M. H., Giangola, J. P., and Balogh, J. (2004). VoiceUser Interface Design. Addison-Wesley.

Cohen, P. R. and Perrault, C. R. (1979). Elements of a plan-based theory of speech acts. Cognitive Science, 3(3), 177–212.

Danieli, M. and Gerbino, E. (1995). Metrics for evaluatingdialogue strategies in a spoken language system. In Pro-ceedings of the 1995 AAAI Spring Symposium on Empir-ical Methods in Discourse Interpretation and Generation,Stanford, CA, pp. 34–39. AAAI Press.

Fikes, R. E. and Nilsson, N. J. (1971). STRIPS: A new ap-proach to the application of theorem proving to problemsolving. Artificial Intelligence, 2, 189–208.

Firth, J. R. (1935). The technique of semantics. Transactionsof the philological society, 34(1), 36–73.

Ginzburg, J. and Sag, I. (2000). Interrogative Investiga-tions: the Form, Meaning and Use of English Interroga-tives. CSLI.

Goldberg, J., Ostendorf, M., and Kirchhoff, K. (2003). Theimpact of response wording in error correction subdialogs.In ISCA Tutorial and Research Workshop on Error Han-dling in Spoken Dialogue Systems.

Goodwin, C. (1996). Transparent vision. In Ochs, E., Sche-gloff, E. A., and Thompson, S. A. (Eds.), Interaction andGrammar, pp. 370–404. Cambridge University Press.

Gravano, A., Hirschberg, J., and Benus, S. (2012). Affirma-tive cue words in task-oriented dialogue. ComputationalLinguistics, 38(1), 1–39.

Grosz, B. J. (1977). The Representation and Use of Focus inDialogue Understanding. Ph.D. thesis, University of Cali-fornia, Berkeley.

Grosz, B. J. and Sidner, C. L. (1980). Plans for discourse. InCohen, P. R., Morgan, J., and Pollack, M. E. (Eds.), Inten-tions in Communication, pp. 417–444. MIT Press.

Hinkelman, E. A. and Allen, J. (1989). Two constraints onspeech act ambiguity. In ACL-89, Vancouver, Canada, pp.212–219.

Hirschberg, J., Litman, D. J., and Swerts, M. (2001). Iden-tifying user corrections automatically in spoken dialoguesystems. In NAACL 2001.

Jefferson, G. (1972). Side sequences. In Sudnow, D. (Ed.),Studies in social interaction, pp. 294–333. Free Press, NewYork.

Jefferson, G. (1984). Notes on a systematic deployment ofthe acknowledgement tokens ‘yeah’ and ‘mm hm’. Papersin Linguistics, 17(2), 197–216.

Jekat, S., Klein, A., Maier, E., Maleck, I., Mast, M., andQuantz, J. (1995). Dialogue acts in verbmobil. Verbmobil–Report–65–95.

Labov, W. and Fanshel, D. (1977). Therapeutic Discourse.Academic Press.

Lemon, O., Georgila, K., Henderson, J., and Stuttle, M.(2006). An ISU dialogue system exhibiting reinforcementlearning of dialogue policies: Generic slot-filling in theTALK in-car system. In EACL-06.

Levesque, H. J., Cohen, P. R., and Nunes, J. H. T. (1990).On acting together. In AAAI-90, Boston, MA, pp. 94–99.Morgan Kaufmann.

Levin, E., Pieraccini, R., and Eckert, W. (2000). A stochas-tic model of human-machine interaction for learning dialogstrategies. IEEE Transactions on Speech and Audio Pro-cessing, 8, 11–23.

Levinson, S. C. (1983). Conversational Analysis, chap. 6.Cambridge University Press.

Levow, G.-A. (1998). Characterizing and recognizing spo-ken corrections in human-computer dialogue. In COLING-ACL, pp. 736–742.

Litman, D. J., Swerts, M., and Hirschberg, J. (2000). Pre-dicting automatic speech recognition performance usingprosodic cues. In NAACL 2000.

Litman, D. J., Walker, M. A., and Kearns, M. (1999). Auto-matic detection of poor speech recognition at the dialoguelevel. In ACL-99, College Park, MA, pp. 309–316.

Page 20: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

20 Chapter 30 • Advanced Dialog Systems

Lochbaum, K. E., Grosz, B. J., and Sidner, C. L. (2000).Discourse structure and intention recognition. In Dale, R.,Moisl, H., and Somers, H. L. (Eds.), Handbook of NaturalLanguage Processing. Marcel Dekker.

Mrksic, N., O’Seaghdha, D., Wen, T.-H., Thomson, B., andYoung, S. (2017). Neural belief tracker: Data-driven dia-logue state tracking. In ACL 2017.

Nagata, M. and Morimoto, T. (1994). First steps toward sta-tistical modeling of dialogue to predict the speech act typeof the next utterance. Speech Communication, 15, 193–203.

Norman, D. A. (1988). The Design of Everyday Things. Ba-sic Books.

Oh, A. H. and Rudnicky, A. I. (2000). Stochastic languagegeneration for spoken dialogue systems. In Proceedingsof the 2000 ANLP/NAACL Workshop on Conversationalsystems-Volume 3, pp. 27–32.

Perrault, C. R. and Allen, J. (1980). A plan-based analysis ofindirect speech acts. American Journal of ComputationalLinguistics, 6(3-4), 167–182.

Purver, M. (2004). The theory and use of clarification re-quests in dialogue. Ph.D. thesis, University of London.

Roy, N., Pineau, J., and Thrun, S. (2000). Spoken dialogmanagement for robots. In ACL-00, Hong Kong.

Russell, S. and Norvig, P. (2002). Artificial Intelligence: AModern Approach (2nd Ed.). Prentice Hall.

Sacks, H., Schegloff, E. A., and Jefferson, G. (1974). Asimplest systematics for the organization of turn-taking forconversation. Language, 50(4), 696–735.

Sag, I. A. and Liberman, M. Y. (1975). The intonational dis-ambiguation of indirect speech acts. In CLS-75, pp. 487–498. University of Chicago.

Schegloff, E. A. (1968). Sequencing in conversational open-ings. American Anthropologist, 70, 1075–1095.

Schegloff, E. A. (1972). Notes on a conversational practice:Formulating place. In Sudnow, D. (Ed.), Studies in socialinteraction, New York. Free Press.

Schegloff, E. A. (1982). Discourse as an interactionalachievement: Some uses of ‘uh huh’ and other things thatcome between sentences. In Tannen, D. (Ed.), AnalyzingDiscourse: Text and Talk, pp. 71–93. Georgetown Univer-sity Press, Washington, D.C.

Shriberg, E., Bates, R., Taylor, P., Stolcke, A., Jurafsky, D.,Ries, K., Coccaro, N., Martin, R., Meteer, M., and Van Ess-Dykema, C. (1998). Can prosody aid the automatic classi-fication of dialog acts in conversational speech?. Languageand Speech (Special Issue on Prosody and Conversation),41(3-4), 439–487.

Singh, S. P., Litman, D. J., Kearns, M., and Walker, M. A.(2002). Optimizing dialogue management with reinforce-ment learning: Experiments with the NJFun system. Jour-nal of Artificial Intelligence Research (JAIR), 16, 105–133.

Stalnaker, R. C. (1978). Assertion. In Cole, P. (Ed.), Prag-matics: Syntax and Semantics Volume 9, pp. 315–332. Aca-demic Press.

Stolcke, A., Ries, K., Coccaro, N., Shriberg, E., Bates, R.,Jurafsky, D., Taylor, P., Martin, R., Meteer, M., and VanEss-Dykema, C. (2000). Dialogue act modeling for au-tomatic tagging and recognition of conversational speech.Computational Linguistics, 26(3), 339–371.

Stoyanchev, S. and Johnston, M. (2015). Localized error de-tection for targeted clarification in a virtual assistant. InICASSP-15, pp. 5241–5245.

Stoyanchev, S., Liu, A., and Hirschberg, J. (2013). Mod-elling human clarification strategies. In SIGDIAL 2013,pp. 137–141.

Stoyanchev, S., Liu, A., and Hirschberg, J. (2014). Towardsnatural clarification questions in dialogue systems. In AISBsymposium on questions, discourse and dialogue.

Sutton, R. S. and Barto, A. G. (1998). Reinforcement Learn-ing: An Introduction. Bradford Books (MIT Press).

Swerts, M., Litman, D. J., and Hirschberg, J. (2000). Cor-rections in spoken dialogue systems. In ICSLP-00, Beijing,China.

Wade, E., Shriberg, E., and Price, P. J. (1992). User behav-iors affecting speech recognition. In ICSLP-92, pp. 995–998.

Walker, M. A. (2000). An application of reinforcement learn-ing to dialogue strategy selection in a spoken dialogue sys-tem for email. Journal of Artificial Intelligence Research,12, 387–416.

Walker, M. A., Fromer, J. C., and Narayanan, S. S. (1998).Learning optimal dialogue strategies: A case study of aspoken dialogue agent for email. In COLING/ACL-98,Montreal, Canada, pp. 1345–1351.

Ward, N. and Tsukahara, W. (2000). Prosodic features whichcue back-channel feedback in English and Japanese. Jour-nal of Pragmatics, 32, 1177–1207.

Weinschenk, S. and Barker, D. T. (2000). Designing Effec-tive Speech Interfaces. Wiley.

Wen, T.-H., Gasic, M., Kim, D., Mrksic, N., Su, P.-H.,Vandyke, D., and Young, S. (2015a). Stochastic lan-guage generation in dialogue using recurrent neural net-works with convolutional sentence reranking. In SIGDIAL2015, pp. 275––284.

Wen, T.-H., Gasic, M., Mrksic, N., Su, P.-H., Vandyke, D.,and Young, S. (2015b). Semantically conditioned lstm-based natural language generation for spoken dialogue sys-tems. In EMNLP 2015.

Wilensky, R. (1983). Planning and Understanding: AComputational Approach to Human Reasoning. Addison-Wesley.

Williams, J., Raux, A., and Henderson, M. (2016). The dia-log state tracking challenge series: A review. Dialogue &Discourse, 7(3), 4–33.

Williams, J. D. and Young, S. J. (2007). Partially observ-able markov decision processes for spoken dialog systems.Computer Speech and Language, 21(1), 393–422.

Wittgenstein, L. (1953). Philosophical Investigations.(Translated by Anscombe, G.E.M.). Blackwell.

Yankelovich, N., Levow, G.-A., and Marx, M. (1995). De-signing SpeechActs: Issues in speech user interfaces. InHuman Factors in Computing Systems: CHI ’95 Confer-ence Proceedings, Denver, CO, pp. 369–376.

Yngve, V. H. (1970). On getting a word in edgewise. InCLS-70, pp. 567–577. University of Chicago.

Page 21: CHAPTER 30 Advanced Dialog Systems - Stanford …jurafsky/slp3/30.pdf · CHAPTER 30Advanced Dialog Systems ... 19*;2 9'&:$/)

Bibliographical and Historical Notes 21

Young, S., Gasic, M., Keizer, S., Mairesse, F., Schatzmann,J., Thomson, B., and Yu, K. (2010). The Hidden Infor-mation State model: A practical framework for POMDP-based spoken dialogue management. Computer Speech &Language, 24(2), 150–174.