• No results found

Distilling dialogues - A method using natural dialogue corpora for dialogue systems development

N/A
N/A
Protected

Academic year: 2021

Share "Distilling dialogues - A method using natural dialogue corpora for dialogue systems development"

Copied!
8
0
0

Loading.... (view fulltext now)

Full text

(1)

D i s t i l l i n g d i a l o g u e s - A m e t h o d u s i n g n a t u r a l d i a l o g u e

d i a l o g u e s y s t e m s d e v e l o p m e n t

A r n e J S n s s o n a n d N i l s D a h l b ~ i c k D e p a r t m e n t o f C o m p u t e r a n d I n f o r m a t i o n S c i e n c e L i n k S p i n g U n i v e r s i t y S - 5 8 1 83, L I N K O P I N G S W E D E N

nilda@ida.liu.se, arnjo@ida.liu.se

c o r p o r a for

A b s t r a c t

We report on a m e t h o d for utilising corpora col- lected in natural settings. It is based on distilling (re-writing) natural dialogues to elicit the type of dialogue t h a t would occur if one the dialogue par- ticipants was a computer instead of a human. T h e m e t h o d is a complement to other means such as Wiz- ard of Oz-studies and un-distilled natural dialogues. We present the distilling m e t h o d and guidelines for distillation. We also illustrate how the m e t h o d af- fects a corpus of dialogues and discuss the pros and cons of three approaches in different phases of dia- logue systems development.

1 I n t r o d u c t i o n

It has been known for quite some time now, t h a t the language used when interacting with a comput- er is different from the one used in dialogues between people, (c.f. JSnsson and Dahlb~ick (1988)). Given t h a t we know t h a t the language will be different, b u t not

how it will be different, we need to base

our development of natural language dialogue sys- tems on a relevant set of dialogue corpora. It is our belief that we need to clarify a number of different issues regarding the collection and use of corpora in the development of speech-only and multimodal dia- logue systems. Exchanging experiences and develop- ing guidelines in this area are as i m p o r t a n t as, and in some sense a necessary pre-requisite to, the develop- ment of computational models of speech, language, and dialogue/discourse. It is interesting to note the difference in the state of art in the field of natu- ral language dialogue systems with t h a t of corpus linguistics, where issues of the usefulness of different samples, the necessary sampling size, representative- ness in corpus design and other have been discussed for quite some time (e.g. (Garside et al., 1997; Atkins et al., 1992; Crowdy, 1993; Biber, 1993)). Also the neighboring area of evaluation of NLP systems (for an overview, see Sparck Jones and Galliers (1996)) seems to have advanced further.

Some work have been done in the area of natu- ral language dialogue systems, e.g. on the design of Wizard of Oz-studies (Dahlb~ck et al., 1998),

on measures for inter-rater reliability (Carletta, 1996), on frameworks for evaluating spoken dialogue agents (Walker et al., 1998) and on the use of differ- ent c o r p o r a in the development of a particular sys- t e m (The Carnegie-Mellon Communicator, Eskenazi et al. (1999)).

T h e question we are addressing in this paper is how to collect and analyse relevant corpora. We be- gin by describing what we consider to be the main advantages and disadvantages of the two currently used methods; studies of h u m a n dialogues and Wiz- ard of Oz-dialogues, especially focusing on the eco- logical validity of the methods. We then describe a m e t h o d called 'distilling dialogues', which can serve as a supplement to the other two.

2 Natural and Wizard of

Oz-Dialogues

T h e advantage of using real dialogues between peo- ple is t h a t t h e y will illustrate which tasks and needs t h a t people actually bring to a p a r t i c u l a r service provider. Thus, on the level of the users' general goals, such dialogues have a high validity. But there are two drawbacks here. First; it is not self-evident t h a t users will have the same task expectations from a c o m p u t e r system as they have with a person. Sec- ond, the language used will differ from the language used when interacting with a computer.

These two disadvantages have been the m a j o r force behind the development of W i z a r d of Oz- methods. T h e advantage here is t h a t the setting will be h u m a n - c o m p u t e r interaction. But there are im- p o r t a n t disadvantages, too. First, on the practical side, the task of setting up a high quality simulation environment and training the operators ('wizards') to use this is a resource consuming task (Dahlb~ck et al., 1998). Second, and probably even more impor- tant, is t h a t we cannot then observe real users using a system for real life tasks, where they bring their own needs, motivations, resources, and constraints to bear. To some extent this problem can be over- come using well-designed so called 'scenarios'. As pointed out in Dahlb~ck (1991), on m a n y levels of analysis the artificiality of the situation will not af-

(2)

fect the language used. An e x a m p l e of this is the p a t t e r n of pronoun-antecedent relations. But since the tasks given to the users are often pre-described by the researchers, this means t h a t this is not a good way of finding out which tasks the users actually want to perform. Nor does it provide a clear enough picture on how the users will act to find something t h a t satisfies their requirements. If e.g. the task is one of finding a charter holiday trip or buying a T V - set within a specified set of constraints (economical and other), it is conceivable t h a t people will stay with the first item t h a t m a t c h e s the specification, whereas in real life they would p r o b a b l y look for alternatives. In our experience, this is primarily a concern if the focus is on the users' goals and plans, b u t is less a p r o b l e m when the interest is on lower- level aspects, such as, syntax or p a t t e r n s of pronoun- antecedent relationship (c.f. Dahlb~ick (1991)).

To summarize; real life dialogues will provide a reasonably correct picture of the way users' ap- proach their tasks, and w h a t tasks they bring to the service provider, but the language used will not give a good a p p r o x i m a t i o n of w h a t the system un- der construction will need to handle. Wizard of Oz- dialogues, on the other hand, will give a reasonable a p p r o x i m a t i o n of some aspects of the language used, but in an artificial context.

T h e usual approach has been to work in three steps. First analyse real h u m a n dialogues, and based on these, in the second phase, design one or m o r e Wizard of Oz-studies. T h e final step is to fine-tune the s y s t e m ' s performance on real users. A good ex- a m p l e of this m e t h o d is presented in Eskenazi et al. (1999). But there are also possible problems with this approach (though we are not claiming t h a t this was the case in their particular project). Eskenazi et al. (1999) asked a h u m a n o p e r a t o r to act ' c o m p u t e r - like' in their Wizard of Oz-phase. T h e advantage is of course t h a t the h u m a n o p e r a t o r will be able to perform all the tasks t h a t is usually provided by this service. T h e disadvantage is t h a t it puts a heavy burden on the h u m a n o p e r a t o r to act as a c o m p u t - er. Since we know t h a t lay-persons' ideas of what computers can and cannot do are in m a n y respects far removed from what is actually the case, we risk introducing some systematic distortion here. And since it is difficult to p e r f o r m consistently in similar situations, we also risk introducing non-systematic distortion here, even in those cases when the 'wiz- a r d ' is an NLP-professional.

Our suggestion is therefore to supplement the above mentioned methods, and bridge the gap be- tween them, by post-processing h u m a n dialogues to give t h e m a computer-like quality. T h e advantage, c o m p a r e d to having people do the simulation on the fly, is b o t h t h a t it can be done with more consis- tency, and also t h a t it can be done by researchers

t h a t actually know what h u m a n - c o m p u t e r n a t u r a l language dialogues can look like. A possible dis- a d v a n t a g e with using b o t h Wizard of Oz-and real c o m p u t e r dialogues, is t h a t users will quickly a d a p t to what the system can provide t h e m with, and will therefore not t r y to use it for tasks t h e y know it cannot perform. Consequently, we will not get a full picture of the different services they would like the system to provide.

A disadvantage with this m e t h o d is, of course, t h a t post-processing takes some time c o m p a r e d to using the n a t u r a l dialogues as they are. There is al- so a concern on the ecological validity of the results, as discussed later.

3

Distilling dialogues

Distilling dialogues, i.e. re-writing h u m a n interac- tions in order to have t h e m reflect what a human- c o m p u t e r interaction could look like involves a num- ber of considerations. T h e m a i n issue is t h a t in cor- p o r a of n a t u r a l dialogues one of the interlocutors is not a dialogue system. The s y s t e m ' s task is instead performed by a h u m a n and the p r o b l e m is how to anticipate the behaviour of a system t h a t does not exist based on the performance of an agent with dif- ferent p e r f o r m a n c e characteristics. One i m p o r t a n t aspect is how to deal with h u m a n features t h a t are not p a r t of what the system is supposed to be a b l e to handle, for instance if the user talks a b o u t things outside of the domain, such as discussing an episode of a recent T V show. It also involves issues on how to handle situations where one of the interlocuters discusses with someone else on a different topic, e.g. discussing the up-coming Friday p a r t y with a friend in the middle of an information providing dialogue with a customer.

It is i m p o r t a n t for the distilling process to have at least an outline of the dialogue system t h a t is under development: Will it for instance have the capacity to recognise users' goals, even if not explicitly stat- ed? Will it be able to reason a b o u t the discourse domain? W h a t services will it provide, and what will be outside its capacity to handle?

In our case, we assume t h a t the planned dialogue system has the ability to reason on various aspects of dialogue and properties of the application. In our current work, and in the examples used for illustra- tion in this paper, we assume a dialogue model t h a t can handle any relevant dialogue phenomenon and also an interpreter and speech recogniser being able to understand any user input t h a t is relevant to the task. T h e r e is is also a powerful domain reason- ing module allowing for more or less any knowledge reasoning on issues t h a t can be accomplished with- in the domain (Flycht-Eriksson, 1999). O u r current system does, however, not have an explicit user task model, as opposed to a system task model (Dahlb~ick

(3)

and JSnsson, 1999), which is included, and thus, we can not assume that the 'system' remembers utter- ances where the user explains its task. Furthermore, as our aim is system development we will not con- sider interaction outside the systems capabilities as relevant to include in the distilled dialogues.

The context of our work is the development a multi-modal dialogue system. However, in our cur- rent work with distilling dialogues, the abilities of a multi-modal system were not fully accounted for. T h e reason for this is t h a t the dialogues would be significantly affected, e.g. a telephone conversation where the user always likes to have the n e x t con- nection, please will result in a table if multi-modal o u t p u t is possible and hence a fair amount of the di- alogne is removed. We have therefore in this paper analysed the corpus assuming a speech-only system, since this is closer to the original telephone conversa- tions, and hence needs fewer assumptions on system performance when distilling the dialogues.

4 D i s t i l l a t i o n g u i d e l i n e s

Distilling dialogues requires guidelines for how to handle various types of utterances. In this section we will present our guidelines for distilling a corpus of telephone conversations between a human infor- mation provider on local buses 1 to be used for devel- oping a multimodal dialogue system (Qvarfordt and JSnsson, 1998; Flycht-Eriksson and JSnsson, 1998; Dahlb~ick et al., 1999; Qvarfordt, 1998). Similar guidelines are used within another project on devel- oping Swedish Dialogue Systems where the domain is travel bureau information.

We can distinguish three types of contributors: 'System' (i.e. a future systems) utterances, User ut- terances, and other types, such as moves by other speakers, and noise.

4.1 Modifying system utterances

T h e problem of modifying 'system' utterances can be divided into two parts: how to change and w h e n to change. T h e y are in some respects intertwined, but as the how-part affects the when-part more we will take this as a starting point.

• The ' s y s t e m ' provides as m u c h relevant infor- m a t i o n as possible at once. This depends on the capabilities of the systems o u t p u t modal- ities. If we have a screen or similar o u t p u t device we present as much as possible which normally is all relevant information. If we, on the other hand, only have spoken o u t p u t the amount of information t h a t the hearer can inter- pret in one utterance must be considered when

1The bus time table dialogues are collected at LinkSping University and are available (in Swedish) on h t t p : / / w w w . i d a . l i u . s e / ~ a r n j o / k f b / d i a l o g e r . h t m l

distilling. The system might in such cases pro- vide less information. T h e principle of provid- ing all relevant information is based on the as- sumption t h a t a computer system often has ac- cess to all relevant information when querying the background system and can also present it more conveniently, especially in a multimodal system (Ahrenberg et al., 1996). A typical ex- ample is the dialogue fragment in figure 1. In this fragment the system provides information on what train to take and how to change to a bus. T h e result of distilling this fragment pro- vides the revised fragment of figure 2. As seen in the fragment of figure 2 we also remove a num- ber of utterances typical for h u m a n interaction, as discussed below.

* S y s t e m u t t e r a n c e s are m a d e m o r e c o m p u t e r - l i k e and do n o t include i r r e l e v a n t i n f o r m a t i o n . T h e latter is seen in $9 in the dialogue in figure 3 where the provided information is not relevant. It could also be possible to remove $5 and re- spond with $7 at once. This, however, depends on if the information grounded in $5-U6 is need- ed for the 'system' in order to know the arrival time or if t h a t could be concluded from U4. This in turn depends on the system's capabili- ties. If we assume t h a t the dialogue system has a model of user tasks, the information in $5-U6 could have been concluded from that. We will, in this case, retain $5-U6 as we do not assume a user task model (Dahlb/ick and JSnsson, 1999) and in order to stay as close to the original di- alogue as possible.

T h e next problem concerns the case when 'system' utterances are changed or removed.

• Dialogue contributions provided by s o m e t h i n g or s o m e o n e other t h a n the u s e r or the ' s y s t e m ' are removed. These are regarded as not being part of the interaction. This means t h a t if some- one interrupts the current interaction, say t h a t the telephone rings during a face-to-face inter- action, the interrupting interaction is normally removed from the corpus.

Furthermore, 'system' interruptions are re- moved. A human can very well interrupt anoth- er h u m a n interlocuter, but a computer system will not do that.

However, this guideline could lead to problems, for instance, when users follow up such interrup- tions. If no information is provided or the in- t e r r u p t e d sequence does not affect the dialogue, we have no problems removing the interruption. T h e problem is what to do when information from the 'system' is used in the continuing dia- logue. For such cases we have no fixed strategy,

(4)

U4: $5: U6: $7: U8: $9: U10: $11: U12: S13: U14: $15:

yes I wonder if you have any m m buses or (.) like express buses leaving from LinkSping to Vadstena (.) on sunday

ja ville undra om ni hade ndgra 5h bussar eUer (.) typ expressbussar sore dkte frdn LinkSping till Vadstena (.) pd sSnda

no the bus does not run on sundays

nej bussen g~r inte pd sSndagar

how can you (.) can you take the train and then change some way (.) because (.) to MjSlby ' n ' so

hur kan man (.) kan man ta tdg d sen byta p~ ndtt sStt (.) fSr de (.) till mjSlby ~ sd

t h a t you can do too yes

de kan du gSra ocksd ja

how (.) do you have any such suggestions

hut (.) har du n~ra n~gra s~na fSrslag

yes let's see (4s) a m o m e n t (15s) now let us see here (.) was it on the sunday you should travel

ja ska se h~ir (4s) eft 5gonblick (15s) nu ska v i s e hSr (.) va de p~ sSndagen du skulle dka pd

yes right afternoon preferably

ja just de eftermidda ggirna

afternoon preferable (.) you have train from LinkSping fourteen twenty nine

eftermidda gSrna (.) du hat t~g frdn LinkSping fjorton d tjugonie

m m m m

and then you will change from MjSlby station six hundred sixty

sd byter du frdn MjSlby station sexhundrasexti

sixhundred sixty

sexhundrasexti

fifteen and ten

femton ~ tie

Figure 1: Dialogue fragment from a real interaction on bus time-table information U4: I wonder if you have any buses or (.) like express buses going from LinkSping

to Vadstena (.) on sunday

S5: no the bus does not run on sundays

U6: how can you (.) can you take the train and then change some way (.) because (.) to MjSlby and so

$7: you can take the train from LinkSping fourteen and twenty nine and then you will change at MjSlby station to bus six hundred sixty at fifteen and ten

Figure 2: A distilled version of the dialogue in figure 1 the dialogue needs to be rearranged depending

on how the information is to be used (c.f. the discussion in the final section of this paper).

• 'System' utterances which are no longer valid are removed. Typical examples of this are the

utterances $7, $9, $11 and $13 in the dialogue fragment of figure 1.

* Remove sequences of utterances where the 'sys- tem' behaves in a way a computer would not do.

For instance jokes, irony, humor, commenting on the other dialogue participant, or dropping the telephone (or whatever is going on in $7

in figure 4). A common case of this is when the ' s y s t e m ' is talking while looking for infor- m a t i o n , $5 in the dialogue fragment of figure 4 is an example of this. Related to this is when the s y s t e m provides its own comments. If we can assume t h a t it has such capabilities they are included, otherwise we remove them.

The system does not repeat information that has already been provided unless explicitly asked to do so. In h u m a n interaction it is not u n c o m m o n

to r e p e a t what has been u t t e r e d for purposes other t h a n to provide grounding information or feedback. This is for instance c o m m o n during

(5)

U4: 'n' I must be at Resecentrum before fourteen and t h i r t y five (.) 'cause we will going to the interstate buses

ja ska va p~ rececentrum innan ]jorton ~ trettifem (.) f5 vi ska till l~ngf~irdsbussarna

$5: a h a (.) ' n ' then you m u s t be there around twenty past two something t h e n

jaha (.) ~ dd behhver du va here strax e~ter tjuge 5vet tvd n~nting d~

U6: yes around t h a t

ja ungefgir

$7: let's see here ( l l s ) two hundred and fourteen R y d end station leaves forty six (.) thirteen ' n ' forty six then you will be down fourteen oh seven (.)

d~ ska v i s e hSr (11s) tv~hundrafjorton Ryd 5ndh~llplatsen gdr ~5rtisex (.) tretton d ]Srtisex d~ dr du nere ~jorton noll sju 5)

U8: a h a

jaha

$9: ' n ' (.) the next one takes you there (.) fourteen t h i r t y seven (.) b u t t h a t is t o o late

(.) ndsta dr du nere 5) ~jorton d trettisju (.) men de 5 ju ~Sr sent

Figure 3: Dialogue fragment from a real interaction on bus time-table information U2: Well, hi (.) I a m going to Ugglegatan eighth

ja hej (.) ja ska till Ugglegatan dtta

$3: Yes

ja

U4: and (.) I wonder (.) it is somewhere in Tannefors

och (.) jag undrar (.) det ligger ndnstans i Tannefors

$5: Yes (.) I will see here one one I will look exactly where it is one m o m e n t please

ja (.) jag ska se hhr eft eft jag ska titta exakt vat det ligger eft 6gonblick barn

U6: Oh Yeah

jar~

$7: (operator disconnects) (25s) m m (.) okey (hs) w h a t the hell (2s) (operator connects again) hello yes

((Telefonisten kopplar ur sig)) (25s) iihh (.) okey (hs) de va sore ]aan (2s) ((Telefonisten kopplar in sig igen)) halld ja

U8: Yes hello

ja hej

$9: It is bus two hundred ten which runs on old tannefors r o a d t h a t you have to take and get off at the bus stop at t h a t bus stop n a m e d v e t e g a t a n

det ~i buss tv~hundratio sore g~r gamla tanne~orsvSgen som du ~ r ~ka ~ g~ av rid den hdllplatsen rid den hdllplatsen sore heter vetegatan.

Figure 4: Dialogue fragment from a n a t u r a l bus t i m e t a b l e interaction search procedures as discussed above.

• The system does not ask for information it has already achieved. For instance asking again if it is on Sunday as in $9 in figure 1. This is not un- c o m m o n in h u m a n interaction and such utter- ances from the user are not removed. However, we can assume t h a t the dialogue system does not forget w h a t has been talked about before. 4.2 M o d i f y i n g u s e r u t t e r a n c e s

T h e general rule is to change user utterances as lit- tle as possible. T h e reason for this is t h a t we do not

want to develop systems where the user needs to restrict h i s / h e r behaviour to the capabilities of the dialogue system. However, there are certain changes m a d e to user utterances, in m o s t cases as a conse- quence of changes of s y s t e m utterances.

Utterances that are no longer valid are removed.

T h e m o s t c o m m o n cases are utterances whose request has already been answered, as seen in the distilled dialogue in figure 2 of the dialogue in figure 1.

(6)

Sl1: sixteen fifty five

sexton ]emti/em

U12: sixteen fifty five (.) a h a

sexton f e m t i / e m (.) jaha

S13: bus line four hundred thirty five

linje ]yrahundra tretti/em

Figure 5: Dialogue f r a g m e n t from a n a t u r a l bus timetable interaction

• Utterances are removed where the user discuss- es things that are in the environment. For instance c o m m e n t i n g the ' s y s t e m s ' clothes or hair. This also includes other types of commu- nicative signals such as laughter based on things outside the interaction, for instance, in the en- vironment of the interlocuters.

• User utterances can also be added in order to make the dialogue continue. In the dialogue in

figure 5 there is nothing in the dialogue explain- ing why the s y s t e m utters S13. In such cases we need to add a user utterance, e.g. Which

bus is that?. However, it might t u r n out t h a t

there are cues, such as intonation, found when listening to the tapes. I f such detailed analyses are carried out, we will, of course, not need to add utterances. Furthermore, it is sometimes the case t h a t the telephone o p e r a t o r deliberate- ly splits the information into chunks t h a t can be comprehended by the user, which then m u s t be considered in the distillation.

5 A p p l y i n g t h e m e t h o d

To illustrate the m e t h o d we will in this section t r y to characterise the results from our distillations. T h e illustration is based on 39 distilled dialogues from the previously mentioned corpus collected with a telephone o p e r a t o r having information on local bus time-tables and persons calling the information ser- vice.

T h e distillation took a b o u t three hours for all 39 dialogues, i.e. it is reasonably fast. T h e distilled dialogues are on the average 27% shorter. However, this varies between the dialogues, at m o s t 73% was removed but there were also seven dialogues t h a t were not changed at all.

At the m o s t 34 utterances where removed from one single dialogue and t h a t was from a dialogue with discussions on where to find a parking lot, i.e. discussions outside the capabilities of the applica- tion. There was one m o r e dialogue where more t h a n 30 utterances were removed and t h a t dialogue is a typical example of dialogues where distillation actu- ally is very useful and also indicates what is normal- ly removed from the dialogues. This particular dia-

logue begins with the user asking for the telephone n u m b e r to ' t h e Lost p r o p e r t y office' for a specific bus operator. However, the o p e r a t o r starts a discussion on what bus the traveller traveled on before provid- ing the requested telephone number. T h e reason for this discussion is p r o b a b l y t h a t the o p e r a t o r knows t h a t different bus companies are utilised and would like to m a k e sure t h a t the user really understands h i s / h e r request. T h e interaction t h a t follows can, thus, in t h a t respect be relevant, but for our pur- pose of developing systems based on an overall goal of providing information, not to understand h u m a n interaction, our dialogue s y s t e m will not able to han- dle such phenomenon (JSnsson, 1996).

T h e dialogues can roughly be divided into five dif- ferent categories based on the users task. T h e dis- cussion in twenty five dialogues were on bus times between various places, often one d e p a r t u r e and one arrival but five dialogues involved more places. In five dialogues the discussion was one price and var- ious types of discounts. Five users wanted to know the telephone n u m b e r to ' t h e Lost p r o p e r t y office', two discussed only bus stops and two discussed how they could utilise their season ticket to travel out- side the trafficking area of the bus company. It is interesting to note t h a t there is no correspondence between the task being performed during the inter- action and the a m o u n t of changes m a d e to the d i a - logue. Thus, if we can assume t h a t the a m o u n t of distillation indicates something a b o u t a user's inter- action style, other factors t h a n the task are impor- t a n t when characterising user behaviour.

Looking at what is altered we find t h a t the most i m p o r t a n t distilling principle is t h a t the ' s y s t e m ' provides all relevant information at once, c.f. fig- ures 1 and 2. This in turn removes utterances pro- vided by b o t h ' s y s t e m ' and user.

Most added utterances, b o t h from the user and the ' s y s t e m ' , provide explicit requests for informa- tion t h a t is later provided in the dialogue, e.g. ut- terance $3 in figure 6. We have added ten utterances in all 39 dialogues, five ' s y s t e m ' utterances and five user utterances. Note, however, t h a t we utilised the transcribed dialogues, without information on into- nation. We would p r o b a b l y not have needed to add this m a n y utterances if we had utilised the tapes. Our reason for not using information on intonation is t h a t we do not assume t h a t our s y s t e m ' s speech recogniser can recognise intonation.

Finally, as discussed above, we did not utilise the full potential of m u l t i - m o d a l i t y when distilling the dialogues. For instance, some dialogues could be further distilled if we had assumed t h a t the system had presented a time-table. One reason for this is t h a t we wanted to c a p t u r e as m a n y interesting as- pects intact as possible. T h e a d v a n t a g e is, thus, t h a t we have a b e t t e r corpus for understanding human-

(7)

U2: Yees hi A n n a Nilsson is my n a m e and I would like to take the bus from R y d center to Resecentrum in LinkSping

jaa hej Anna Nilsson heter jag och jag rill ~ka buss ~r~n Ryds centrum till resecentrum i LinkSping.

$3: m m W h e n d o y o u w a n t t o l e a v e ?

mm N~ir r i l l d u £ k a ?

U4: 'n' I must be at Resecentrum before fourteen and t h i r t y five (.) 'cause we will going to the interstate buses

ja ska va p~ rececentrum innan fjorton d trettifem (.) f5 vi ska till l~ngfiirdsbussarna

Figure 6: Distilled dialogue f r a g m e n t with added u t t e r a n c e c o m p u t e r interaction and can from t h a t corpus do

a second distillation where we focus more on multi- modal interaction.

6

D i s c u s s i o n

We have been presenting a m e t h o d for distilling hu- m a n dialogues to m a k e t h e m resemble h u m a n com- puter interaction, in order to utilise such dialogues as a knowledge source when developing dialogue sys- tems. Our own m a i n purpose has been to use t h e m for developing m u l t i m o d a l systems, however, as dis- cussed above, we have in this p a p e r rather assumed a speech-only system. B u t we believe t h a t the basic approach can be used also for m u l t i - m o d a l systems and other kinds of n a t u r a l language dialogue sys- tems.

It is i m p o r t a n t to be aware of the limitations of the method, and how 'realistic' the produced result will be, c o m p a r e d to a dialogue with the final sys- tem. Since we are changing the dialogue moves, by for instance providing all required information in one move, or never asking to be reminded of what the us- er has previously requested, it is obvious t h a t w h a t follows after the changed sequence would p r o b a b l y be affected one way or another. A consequence of this is t h a t the resulting dialogue is less accurate as a model of the entire dialogue. It is therefore not an ideal candidate for trying out the systems over-all performance during s y s t e m development. B u t for the smaller sub-segments or sub-dialogues, we be- lieve t h a t it creates a good a p p r o x i m a t i o n of w h a t will take place once the system is up and running. Furthermore, we believe distilled dialogues in some respects to be m o r e realistic t h a n Wizard of Oz- dialogues collected with a wizard acting as a com- puter.

Another issue, t h a t has been discussed previously in the description of the m e t h o d , is t h a t the distilling is m a d e based on a particular view of what a dialogue with a c o m p u t e r will look like. While not necessari- ly being a detailed and specific model, it is at least an instance of a class of c o m p u t e r dialogue models.

One example of this is w h e t h e r the s y s t e m is m e a n t to acquire information on t h e user's underlying mo- tivations or goals or not. In the examples presented, we have not assumed such capabilities, b u t this as- s u m p t i o n is not an absolute necessity. We believe, however, t h a t the distilling process should be based on one such model, not the least to ensure a con- sistent t r e a t m e n t of similar recurring p h e n o m e n a at different places in the corpora.

T h e validity of the results based on analysing dis- tilled dialogues depends p a r t l y on how the distilla- tion has been carried out. Even when using n a t u r a l dialogues we can have situations where the interac- tion is somewhat mysterious, for instance, if some of t h e dialogue participants behaves irrational such as not providing feedback or being t o o elliptical. How- ever, if careful considerations have been m a d e to stay as close to the original dialogues as possible, we be- lieve t h a t distilled dialogues will reflect w h a t a hu- m a n would consider to be a n a t u r a l interaction. A c k n o w l e d g m e n t s

This work results from a n u m b e r of projects on de- velopment of n a t u r a l language interfaces s u p p o r t e d by T h e Swedish T r a n s p o r t & C o m m u n i c a t i o n s Re- search Board (KFB) and the joint Research P r o g r a m for Language Technology ( H S F R / N U T E K ) . We are indebted to the p a r t i c i p a n t s of the Swedish Dialogue Systems project, especially to Staffan Larsson, Lena S a n t a m a r t a , and Annika Flycht-Eriksson for inter- esting discussions on this topic.

R e f e r e n c e s

Lars Ahrenberg, Nils Dahlb~ck, Arne JSnsson, and /~ke Thur~e. 1996. Customizing interac- tion for natural language interfaces. LinkSpin9 Electronic articles in Computer and Informa- tion Science, also in Notes from Workshop on Pragmatics in Dialogue, The XIV:th Scandi- navian Conference of Linguistics and the VI- II:th Conference of Nordic and General Linguis-

(8)

tics, GSteborg, Sweden, 1993,

1(1), October, 1.

http : / / www.ep.liu.se / ea / cis /1996 / O01/.

Sue Atkins, Jeremy Clear, and Nicholas Ostler.

1992. Corpus design criteria.

Literary and Lin-

guistic Computing,

7(1):1-16.

Douglas Biber. 1993. Representativeness in cor-

pus design.

Literary and Linguistic Computing,

8(4):244-257.

Jean Carletta. 1996. Assessing agreement on classi-

fication tasks: The kappa statistic.

Computation-

al Linguistics,

22(2):249-254.

Steve Crowdy. 1993. Spoken corpus design.

Literary

and Linguistic Computing,

8(4):259-265.

Nils Dahlb/ick and Arne JSnsson. 1999. Knowledge

sources in spoken dialogue systems. In

Proceed-

ings of Eurospeech'99, Budapest, Hungary.

Nils Dahlb/ick, Arne JSnsson, and Lars Ahrenberg. 1998. Wizard of oz studies - why and how. In Mark Maybury & Wolfgang Wahlster, editor,

Readings in Intelligent User Interfaces.

Morgan Kaufmann.

Ntis Dahlb/ick, Annika Flycht-Eriksson, Arne

JSnsson, and Pernilla Qvarfordt. 1999. An ar- chitecture for multi-modal natural dialogue sys-

tems. In

Proceedings of ESCA Tutorial and Re-

search Workshop (ETRW) on Interactive Dialogue

in Multi-Modal Systems, Germany.

Nils Dahlb/ick. 1991.

Representations of Discourse,

Cognitive and Computational Aspects.

Ph.D. the- sis, LinkSping University.

Maxine Eskenazi, Alexander Rudnicki, Karin Grego- ry, Paul Constantinides, Robert Brennan, Christi- na Bennett, and Jwan Allen. 1999. Data collec- tion and processing in the carnegie mellon com-

municator. In

Proceedings of Eurospeech'99, Bu-

dapest, Hungary.

Annika Flycht-Eriksson and Arne JSnsson. 1998. A spoken dialogue system utilizing spatial informa-

tion. In

Proceedings of ICSLP'98, Sydney, Aus-

tralia.

Annika Flycht-Eriksson. 1999. A survey of knowl-

edge sources in dialogue systems. In

Proceedings

of lJCAI-99 Workshop on Knowledge and Reason-

ing in Practical Dialogue Systems, August, Stock-

holm.

Roger Garside, Geoffrey Leech, and Anthony

MeEnery. 1997.

Corpus Annotation.

Longman.

Arne JSnsson and Nils Dahlb/ick. 1988. Talking to a computer is not like talking to your best friend. In

Proceedings of the First Scandinavian Conference

on Artificial InterUigence, Tvoms¢.

Arne JSnsson. 1996. Natural language generation

without intentions. In

Proceedings of ECAI'96

Workshop Gaps and Bridges: New Directions

in Planning and Natural Language Generation,

pages 102-104.

Pernilla Qvarfordt and Arne JSnsson. 1998. Effects of using speech in timetable information systems

for www. In

Proceedings of ICSLP'98, Sydney,

Australia.

Pernilla Qvarfordt. 1998. Usability of multimodal

timetables: Effects of different levels of do-

main knowledge on usability. Master's thesis, LinkSping University.

Karen Sparck Jones and Julia R. Galliers. 1996.

Evaluating Natural Language Processing Systems.

Springer Verlag.

Marilyn A. Walker, Diane J. Litman, Candace A. Kamm, and Alicia Abella. 1998. Paradise: A framework for evaluating spoken dialogue agents. In Mark Maybury & Wolfgang Wahlster, editor,

Readings in Intelligent User Interfaces.

Morgan Kaufmann.

References

Related documents

6.1.1 The Reasons for Choosing HiQ for a Case Study. 2001 can so far be characterised as a dramatic year for Swedish IT companies. Huge financial problems have caused a vast number

A listening test was conducted where 23 subjects rated the variation, naturalness and artifacts of a recorded spoken phrase where three altered versions manipulated by altering

N N Student apartment 47m2 Student apartment 47m2 Student apartment 47m2 Common space Common space Reser ved for gar dening Reser ved for gar dening Reser ved for gar dening

För omsättning till den här undersökningen och principen överraskning ser refe- renshypotesen ut på följande sätt: Överraskning = vilseledning + operationssäkerhet + tempo

granskning av Evans egen text - som här i sin förekomst är ett historiskt faktum - är begreppet ”historiskt faktum/historiska fakta”.. Tveksamma som historiska fakta är

Noting that some studies suggest a possible increase in storm activity, over e.g.; the North Atlantic (IPCC, 2013 ); and the North Sea (Lindner & Rummukainen, 2013 ), and in

I SÖ-sluttningen ned mot Lena-dalen finns ett rätt trivialt bestånd av löv som längst i SV består av rätt klen ekskog och längre upp i NÖ blandskog av sälg, björk, asp och

Intressant är att informanterna arbetar på olika sätt med papper, inte bara sett till teknik utan även till hur de får in pengar på sitt arbete. Det handlar inte