screenplay summarization using latent narrative structure · 2020. 6. 20. · screenplay...

14
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 1920–1933 July 5 - 10, 2020. c 2020 Association for Computational Linguistics 1920 Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi 1 Frank Keller 1 Lea Frermann 2 Mirella Lapata 1 1 Institute for Language, Cognition and Computation School of Informatics, University of Edinburgh 2 School of Computing and Information Systems University of Melbourne [email protected], [email protected], [email protected], [email protected] Abstract Most general-purpose extractive summariza- tion models are trained on news articles, which are short and present all important information upfront. As a result, such models are biased by position and often perform a smart selection of sentences from the beginning of the doc- ument. When summarizing long narratives, which have complex structure and present in- formation piecemeal, simple position heuris- tics are not sufficient. In this paper, we pro- pose to explicitly incorporate the underlying structure of narratives into general unsuper- vised and supervised extractive summarization models. We formalize narrative structure in terms of key narrative events (turning points) and treat it as latent in order to summarize screenplays (i.e., extract an optimal sequence of scenes). Experimental results on the CSI corpus of TV screenplays, which we augment with scene-level summarization labels, show that latent turning points correlate with im- portant aspects of a CSI episode and improve summarization performance over general ex- tractive algorithms, leading to more complete and diverse summaries. 1 Introduction Automatic summarization has enjoyed renewed interest in recent years thanks to the popular- ity of modern neural network-based approaches (Cheng and Lapata, 2016; Nallapati et al., 2016, 2017; Zheng and Lapata, 2019) and the avail- ability of large-scale datasets containing hundreds of thousands of document–summary pairs (Sand- haus, 2008; Hermann et al., 2015; Grusky et al., 2018; Narayan et al., 2018; Fabbri et al., 2019; Liu and Lapata, 2019). Most efforts to date have con- centrated on the summarization of news articles which tend to be relatively short and formulaic following an “inverted pyramid” structure which places the most essential, novel and interesting el- Victim: Mike Kimble, found in a Body Farm. Died 6 hours ago, unknown cause of death. &6, GLVFRYHU FRZ WLVVXH LQ 0LNHV ERG\¬ Cross-contamination is suggested. Probable cause of death: Mike's house has been set on fire. CSI finds blood: Mike was murdered, fire was a cover up. First suspects: Mike's fiance, Jane and her ex-husband, Russ. &6, ÀQGV SKRWRV LQ 0LNHV KRXVH RI -DQHV GDXJKWHU -RGLH SRVLQJ QDNHG Mike is now a suspect of abusing Jodie. Russ allows CSI to examine his gun. &6, GLVFRYHUV WKDW WKH EXOOHW WKDW NLOOHG 0LNH ZDV PDGH RI IUR]HQ EHHI WKDW PHOW LQVLGH KLP 7KH\ DOVR ÀQG EHHI LQ 5XVV JXQ Russ confesses that he knew that Mike was abusing Jody, so he confronted and killed him. CSI discovers that the naked photos were taken on a boat, which belongs to Russ. &6, GLVFRYHUV WKDW LW ZDV 5XVV ZKR ZDV DEXVLQJ KLV GDXJKWHU EDVHG RQ ÁXLGV IRXQG LQ KLV VOHHSLQJ EDJ DQG ODWHU NLOOHG 0LNH ZKR WULHG WR KHOS -RGLH 5XVV LV JLYHQ EDLO VLQFH QR MXU\ ZRXOG FRQYLFW D SURWHFWLYH IDWKHU Russ receives a mandatory life sentence. 6HWXS 1HZ 6LWXDWLRQ 3URJUHVV &RPSOLFDWLRQV 7KH ÀQDO SXVK $IWHUPDWK 2SSRUWXQLW\ &KDQJH RI 3ODQV 3RLQW RI QR 5HWXUQ 0DMRU 6HWEDFN &OLPD[ Figure 1: Example of narrative structure for episode “Burden of Proof” from TV series Crime Scene Inves- tigation (CSI); turning points are highlighted in color. ements of a story in the beginning and support- ing material and secondary details afterwards. The rigid structure of news articles is expedient since important passages can be identified in predictable locations (e.g., by performing a “smart selection” of sentences from the beginning of the document) and the structure itself can be explicitly taken into account in model design (e.g., by encoding the rel- ative and absolute position of each sentence). In this paper we are interested in summarizing longer narratives, i.e., screenplays, whose form and structure is far removed from newspaper ar- ticles. Screenplays are typically between 110 and 120 pages long (20k words), their content is bro- ken down into scenes, which contain mostly dia- logue (lines the actors speak) as well as descrip- tions explaining what the camera sees. Moreover, screenplays are characterized by an underlying narrative structure, a sequence of events by which

Upload: others

Post on 05-Sep-2020

12 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 1920–1933July 5 - 10, 2020. c©2020 Association for Computational Linguistics

1920

Screenplay Summarization Using Latent Narrative Structure

Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2 Mirella Lapata1

1Institute for Language, Cognition and ComputationSchool of Informatics, University of Edinburgh

2School of Computing and Information SystemsUniversity of Melbourne

[email protected], [email protected],[email protected], [email protected]

Abstract

Most general-purpose extractive summariza-tion models are trained on news articles, whichare short and present all important informationupfront. As a result, such models are biased byposition and often perform a smart selectionof sentences from the beginning of the doc-ument. When summarizing long narratives,which have complex structure and present in-formation piecemeal, simple position heuris-tics are not sufficient. In this paper, we pro-pose to explicitly incorporate the underlyingstructure of narratives into general unsuper-vised and supervised extractive summarizationmodels. We formalize narrative structure interms of key narrative events (turning points)and treat it as latent in order to summarizescreenplays (i.e., extract an optimal sequenceof scenes). Experimental results on the CSIcorpus of TV screenplays, which we augmentwith scene-level summarization labels, showthat latent turning points correlate with im-portant aspects of a CSI episode and improvesummarization performance over general ex-tractive algorithms, leading to more completeand diverse summaries.

1 Introduction

Automatic summarization has enjoyed renewedinterest in recent years thanks to the popular-ity of modern neural network-based approaches(Cheng and Lapata, 2016; Nallapati et al., 2016,2017; Zheng and Lapata, 2019) and the avail-ability of large-scale datasets containing hundredsof thousands of document–summary pairs (Sand-haus, 2008; Hermann et al., 2015; Grusky et al.,2018; Narayan et al., 2018; Fabbri et al., 2019; Liuand Lapata, 2019). Most efforts to date have con-centrated on the summarization of news articleswhich tend to be relatively short and formulaicfollowing an “inverted pyramid” structure whichplaces the most essential, novel and interesting el-

Victim: Mike Kimble, found in a Body Farm. Died 6hours ago, unknown cause of death.CSI discover cow tissue in Mike's body. Cross-contamination is suggested. Probablecause of death: Mike's house has been set onfire. CSI finds blood: Mike was murdered, fire wasa cover up. First suspects: Mike's fiance, Janeand her ex-husband, Russ. CSI finds photos in Mike's house of Jane'sdaughter, Jodie, posing naked.Mike is now a suspect of abusing Jodie. Russallows CSI to examine his gun.CSI discovers that the bullet that killed Mikewas made of frozen beef that melt inside him.They also find beef in Russ' gun.Russ confesses that he knew that Mike wasabusing Jody, so he confronted and killed him.

CSI discovers that the naked photos were takenon a boat, which belongs to Russ.CSI discovers that it was Russ who wasabusing his daughter based on fluids found inhis sleeping bag and later killed Mike who triedto help Jodie.

Russ is given bail, since no jury would convicta protective father.

Russ receives a mandatory life sentence.

Setup

NewSituation

Progress

Complications

The final push

Aftermath

Opportunity

Change ofPlans

Point ofno Return

MajorSetback

Climax

Figure 1: Example of narrative structure for episode“Burden of Proof” from TV series Crime Scene Inves-tigation (CSI); turning points are highlighted in color.

ements of a story in the beginning and support-ing material and secondary details afterwards. Therigid structure of news articles is expedient sinceimportant passages can be identified in predictablelocations (e.g., by performing a “smart selection”of sentences from the beginning of the document)and the structure itself can be explicitly taken intoaccount in model design (e.g., by encoding the rel-ative and absolute position of each sentence).

In this paper we are interested in summarizinglonger narratives, i.e., screenplays, whose formand structure is far removed from newspaper ar-ticles. Screenplays are typically between 110 and120 pages long (20k words), their content is bro-ken down into scenes, which contain mostly dia-logue (lines the actors speak) as well as descrip-tions explaining what the camera sees. Moreover,screenplays are characterized by an underlyingnarrative structure, a sequence of events by which

Page 2: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

1921

Screenplay Latent Narrative Structure

TP1: Introduction

TP3: Commitment

TP2: Goal definition 

TP4: Setback

TP5: Ending

Summary scenes Video summaryrelevant to TP2

relevant to TP5

irrelevant

Figure 2: We first identify scenes that act as turningpoints (i.e., key events that segment the story into sec-tions). We next create a summary by selecting informa-tive scenes, i.e.,semantically related to turning points.

a story is defined (Cutting, 2016), and by thestory’s characters and their roles (Propp, 1968).Contrary to news articles, the gist of the story in ascreenplay is not disclosed at the start, informationis often revealed piecemeal; characters evolve andtheir actions might seem more or less importantover the course of the narrative. From a modelingperspective, obtaining training data is particularlyproblematic: even if one could assemble screen-plays and corresponding summaries (e.g., by min-ing IMDb or Wikipedia), the size of such a corpuswould be at best in the range of a few hundredexamples not hundreds of thousands. Also notethat genre differences might render transfer learn-ing (Pan and Yang, 2010) difficult, e.g., a modeltrained on movie screenplays might not generalizeto sitcoms or soap operas.

Given the above challenges, we introduce anumber of assumptions to make the task feasible.Firstly, our goal is to produce informative sum-maries, which serve as a surrogate to reading thefull script or watching the entire film. Secondly,we follow Gorinski and Lapata (2015) in con-ceptualizing screenplay summarization as the taskof identifying a sequence of informative scenes.Thirdly, we focus on summarizing television pro-grams such as CSI: Crime Scene Investigation (Fr-

ermann et al., 2018) which revolves around a teamof forensic investigators solving criminal cases.Such programs have a complex but well-definedstructure: they open with a crime, the crime sceneis examined, the victim is identified, suspects areintroduced, forensic clues are gathered, suspectsare investigated, and finally the case is solved.

In this work, we adapt general-purpose extrac-tive summarization algorithms (Nallapati et al.,2017; Zheng and Lapata, 2019) to identify infor-mative scenes in screenplays and instill in themknowledge about narrative film structure (Hauge,2017; Cutting, 2016; Freytag, 1896). Specifically,we adopt a scheme commonly used by screen-writers as a practical guide for producing success-ful screenplays. According to this scheme, well-structured stories consist of six basic stages whichare defined by five turning points (TPs), i.e., eventswhich change the direction of the narrative, anddetermine the story’s progression and basic the-matic units. In Figure 1, TPs are highlighted fora CSI episode. Although the link between turningpoints and summarization has not been previouslymade, earlier work has emphasized the importanceof narrative structure for summarizing books (Mi-halcea and Ceylan, 2007) and social media content(Kim and Monroy-Hernandez, 2015). More re-cently, Papalampidi et al. (2019) have shown howto identify turning points in feature-length screen-plays by projecting synopsis-level annotations.

Crucially, our method does not involve man-ually annotating turning points in CSI episodes.Instead, we approximate narrative structure au-tomatically by pretraining on the annotations ofthe TRIPOD dataset of Papalampidi et al. (2019)and employing a variant of their model. We findthat narrative structure representations learned ontheir dataset (which was created for feature-lengthfilms), transfer well across cinematic genres andcomputational tasks. We propose a framework forend-to-end training in which narrative structure istreated as a latent variable for summarization. Weextend the CSI dataset (Frermann et al., 2018) withbinary labels indicating whether a scene should beincluded in the summary and present experimentswith both supervised and unsupervised summa-rization models. An overview of our approach isshown in Figure 2.

Our contributions can be summarized as fol-lows: (a) we develop methods for instilling knowl-edge about narrative structure into generic su-

Page 3: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

1922

pervised and unsupervised summarization algo-rithms; (b) we provide a new layer of annotationsfor the CSI corpus, which can be used for researchin long-form summarization; and (c) we demon-strate that narrative structure can facilitate screen-play summarization; our analysis shows that keyevents identified in the latent space correlate withimportant summary content.

2 Related Work

A large body of previous work has focused on thecomputational analysis of narratives (Mani, 2012;Richards et al., 2009). Attempts to analyze howstories are written have been based on sequencesof events (Schank and Abelson, 1975; Chambersand Jurafsky, 2009), plot units (McIntyre and Lap-ata, 2010; Goyal et al., 2010; Finlayson, 2012) andtheir structure (Lehnert, 1981; Rumelhart, 1980),as well as on characters or personas in a narrative(Black and Wilensky, 1979; Propp, 1968; Bam-man et al., 2014, 2013; Valls-Vargas et al., 2014)and their relationships (Elson et al., 2010; Agarwalet al., 2014; Srivastava et al., 2016).

As mentioned earlier, work on summarizationof narratives has had limited appeal, possibly dueto the lack of annotated data for modeling andevaluation. Kazantseva and Szpakowicz (2010)summarize short stories based on importance cri-teria (e.g., whether a segment contains protagonistor location information); they create summaries tohelp readers decide whether they are interested inreading the whole story, without revealing its plot.Mihalcea and Ceylan (2007) summarize bookswith an unsupervised graph-based approach op-erating over segments (i.e., topical units). Theiralgorithm first generates a summary for each seg-ment and then an overall summary by collectingsentences from the individual segment summaries.

Focusing on screenplays, Gorinski and Lapata(2015) generate a summary by extracting an opti-mal chain of scenes via a graph-based approachcentered around the main characters. In a sim-ilar fashion, Tsoneva et al. (2007) create videosummaries for TV series episodes; their algorithmranks sub-scenes in terms of importance using fea-tures based on character graphs and textual cuesavailable in the subtitles and movie scripts. Vicolet al. (2018) introduce the MovieGraphs dataset,which also uses character-centered graphs to de-scribe the content of movie video clips.

Our work synthesizes various strands of re-

search on narrative structure analysis (Cutting,2016; Hauge, 2017), screenplay summarization(Gorinski and Lapata, 2015), and neural networkmodeling (Dong, 2018). We focus on extractivesummarization and our goal is to identify an op-timal sequence of key events in a narrative. Weaim to create summaries which re-tell the plot of astory in a concise manner. Inspired by recent neu-ral network-based approaches (Cheng and Lapata,2016; Nallapati et al., 2017; Zhou et al., 2018;Zheng and Lapata, 2019), we develop supervisedand unsupervised models for our summarizationtask based on neural representations of scenesand how these relate to the screenplay’s narra-tive structure. Contrary to most previous workwhich has focused on characters, we select sum-mary scenes based on events and their importancein the story. Our definition of narrative structureclosely follows Papalampidi et al. (2019). How-ever, the model architectures we propose are gen-eral and could be adapted to different plot analysisschemes (Field, 2005; Vogler, 2007). To overcomethe difficulties in evaluating summaries for longernarratives, we also release a corpus of screenplayswith scenes labeled as important (summary wor-thy). Our annotations augment an existing datasetbased on CSI episodes (Frermann et al., 2018),which was originally developed for incrementalnatural language understanding.

3 Problem Formulation

Let D denote a screenplay consisting of a se-quence of scenes D = {s1,s2, . . . ,sn}. Our aim isto select a subset D ′ = {si, . . . ,sk} consisting ofthe most informative scenes (where k < n). Notethat this definition produces extractive summaries;we further assume that selected scenes are pre-sented according to their order in the screenplay.We next discuss how summaries can be created us-ing both unsupervised and supervised approaches,and then move on to explain how these are adaptedto incorporate narrative structure.

3.1 Unsupervised Screenplay Summarization

Our unsupervised model is based on an extensionof TEXTRANK (Mihalcea and Tarau, 2004; Zhengand Lapata, 2019), a well-known algorithm for ex-tractive single-document summarization. In oursetting, a screenplay is represented as a graph, inwhich nodes correspond to scenes and edges be-tween scenes si and s j are weighted by their simi-

Page 4: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

1923

larity ei j. A node’s centrality (importance) is mea-sured by computing its degree:

centrality(si) = λ1 ∑j<i

ei j +λ2 ∑j>i

ei j (1)

where λ1+λ2 = 1. The modification introduced inZheng and Lapata (2019) takes directed edges intoaccount, capturing the intuition that the centralityof any two nodes is influenced by their relative po-sition. Also note that the edges of preceding andfollowing scenes are differentially weighted by λ1and λ2.

Although earlier implementations of TEXT-RANK (Mihalcea and Tarau, 2004) compute nodesimilarity based on symbolic representations suchas tf*idf, we adopt a neural approach. Specifically,we obtain sentence representations based on a pre-trained encoder. In our experiments, we rely onthe Universal Sentence Encoder (USE; Cer et al.2018), however, other embeddings are possible.1

We represent a scene by the mean of its sentencerepresentations and measure scene similarity ei j

using cosine.2 As in the original TEXTRANK al-gorithm (Mihalcea and Tarau, 2004), scenes areranked based on their centrality and the M mostcentral ones are selected to appear in the summary.

3.2 Supervised Screenplay Summarization

Most extractive models frame summarization asa classification problem. Following a recent ap-proach (SUMMARUNNER; Nallapati et al. 2017),we use a neural network-based encoder to buildrepresentations for scenes and apply a binary clas-sifier over these to predict whether they shouldbe in the summary. For each scene si ∈ D , wepredict a label yi ∈ {0,1} (where 1 means thatsi must be in the summary) and assign a scorep(yi|si,D,θ) quantifying si’s relevance to the sum-mary (θ denotes model parameters). We assem-ble a summary by selecting M sentences with thetop p(1|si,D,θ).

We calculate sentence representations via thepre-trained USE encoder (Cer et al., 2018); a sceneis represented as the weighted sum of the repre-sentations of its sentences, which we obtain froma BiLSTM equipped with an attention mechanism.Next, we compute richer scene representations bymodeling surrounding context of a given scene.

1USE performed better than BERT in our experiments.2We found cosine to be particularly effective with USE

representations; other metrics are also possible.

We encode the screenplay with a BiLSTM net-work and obtain contextualized representations s′ifor scenes si by concatenating the hidden layers ofthe forward

−→hi and backward

←−hi LSTM, respec-

tively: s′i = [−→hi ;←−hi ]. The vector s′i therefore repre-

sents the content of the ith scene.We also estimate the salience of scene si by

measuring its similarity with a global screenplaycontent representation d. The latter is the weightedsum of all scene representations s1,s2, . . . ,sn. Wecalculate the semantic similarity between s′i and dby computing the element-wise dot product bi, co-sine similarity ci, and pairwise distance ui betweentheir respective vectors:

bi = s′i�d ci =s′i ·d∥∥s′i∥∥‖d‖ (2)

ui =s′i ·d

max(‖s′i‖2 · ‖d‖2)(3)

The salience vi of scene si is the concatenation ofthe similarity metrics: vi = [bi;ci;ui]. The contentvector s′i and the salience vector vi are concate-nated and fed to a single neuron that outputs theprobability of a scene belonging to the summary.3

3.3 Narrative Structure

We now explain how to inject knowledge aboutnarrative structure into our summarization models.For both models, such knowledge is transferredvia a network pre-trained on the TRIPOD4 datasetintroduced by Papalampidi et al. (2019). Thisdataset contains 99 movies annotated with turningpoints. TPs are key events in a narrative that definethe progression of the plot and occur between con-secutive acts (thematic units). It is often assumed(Cutting, 2016) that there are six acts in a film(Figure 1), each delineated by a turning point (ar-rows in the figure). Each of the five TPs has also awell-defined function in the narrative: we presenteach TP alongside with its definition as stated inscreenwriting theory (Hauge, 2017) and adoptedby Papalampidi et al. (2019) in Table 1 (see Ap-pendix A for a more detailed description of nar-rative structure theory). Papalampidi et al. (2019)identify scenes in movies that correspond to thesekey events as a means for analyzing the narrative

3Aside from salience and content, Nallapati et al. (2017)take into account novelty and position-related features. Weignore these as they are specific to news articles and denotethe modified model as SUMMARUNNER*.

4https://github.com/ppapalampidi/TRIPOD

Page 5: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

1924

Turning Point Definition

TP1: OpportunityIntroductory event that occurs afterthe presentation of the story setting.

TP2: Change of PlansEvent where the main goal of thestory is defined.

TP3: Point of No ReturnEvent that pushes the main charac-ter(s) to fully commit to their goal.

TP4: Major SetbackEvent where everything falls apart(temporarily or permanently).

TP5: ClimaxFinal event of the main story, mo-ment of resolution.

Table 1: Turning points and their definitions as givenin Papalampidi et al. (2019)

structure of movies. They collect sentence-levelTP annotations for plot synopses and subsequentlyproject them via distant supervision onto screen-plays, thereby creating silver-standard labels. Weutilize this silver-standard dataset in order to pre-train a network which performs TP identification.

TP Identification Network We first encodescreenplay scenes via a BiLSTM equipped with anattention mechanism. We then contextualize themwith respect to the whole screenplay via a secondBiLSTM. Next, we compute topic-aware scenerepresentations ti via a context interaction layer(CIL) as proposed in Papalampidi et al. (2019).CIL is inspired by traditional segmentation ap-proaches (Hearst, 1997) and measures the seman-tic similarity of the current scene with a precedingand following context window in the screenplay.Hence, the topic-aware scene representations alsoencode the degree to which each scene acts as atopic boundary in the screenplay.

In the final layer, we employ TP-specific atten-tion mechanisms to compute the probability pi j

that scene ti represents the jth TP in the screen-play. Note that we expect the TP-specific atten-tion distributions to be sparse, as there are onlya few scenes which are relevant for a TP (recallthat TPs are boundary scenes between sections).To encourage sparsity, we add a low temperaturevalue τ (Hinton et al., 2015) to the softmax part ofthe attention mechanisms:

gi j = tanh(Wjti +b j), g j ∈ [−1,1] (4)

pi j =exp(gi j/τ)

∑Tt=1 exp(gt j/τ)

,T

∑i=1

pi j = 1 (5)

where Wj,b j represent the trainable weights of theattention layer of the jth TP.

Unsupervised SUMMER We now introduceour model, SUMMER (short for Screenplay

Summarization with Narrative Structure).5 Wefirst present an unsupervised variant which mod-ifies the computation of scene centrality in the di-rected version of TEXTRANK (Equation (1)).

Specifically, we use the pre-trained network de-scribed in Section 3.3 to obtain TP-specific at-tention distributions. We then select an overallscore fi for each scene (denoting how likely it isto act as a TP). We set fi = max j∈[1,5] pi j, i.e., tothe pi j value that is highest across TPs. We incor-porate these scores into centrality as follows:

centrality(si)=λ1∑j<i(ei j+ f j)+λ2∑

j>i(ei j+ fi) (6)

Intuitively, we add the f j term in the forward sumin order to incrementally increase the centralityscores of scenes as the story moves on and we en-counter more TP events (i.e., we move to later sec-tions in the narrative). At the same time, we addthe fi term in the backward sum in order to alsoincrease the scores of scenes identified as TPs.

Supervised SUMMER We also propose a su-pervised variant of SUMMER following the basicmodel formulation in Section 3.3. We still repre-sent a scene as the concatenation of a content vec-tor s′ and salience vector v′, which serve as inputto a binary classifier. However, we now modifyhow salience is determined; instead of comput-ing a general global content representation d forthe screenplay, we identify a sequence of TPs andmeasure the semantic similarity of each scene withthis sequence. Our model is depicted in Figure 3.

We utilize the pre-trained TP network (Fig-ures 3(a) and (b)) to compute sparse attentionscores over scenes. In the supervised setting,where gold-standard binary labels provide a train-ing signal, we fine-tune the network in an end-to-end fashion on summarization (Figure 3(c)). Wecompute the TP representations via the attentionscores; we calculate a vector t p j as the weightedsum of all topic-aware scene representations t pro-duced via CIL: t p j = ∑i∈[1,N] pi jti, where N is thenumber of scenes in a screenplay. In practice, onlya few scenes contribute to t p j due to the τ param-eter in the softmax function (Equation (5)).

A TP-scene interaction layer measures the se-mantic similarity between scenes ti and latent TPrepresentations t p j (Figure 3(c)). Intuitively, acomplete summary should contain scenes which

5We make our code publicly available at https://github.com/ppapalampidi/SUMMER.

Page 6: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

1925

s1 sk sM. . .. . .

Screenplay Encoder(BiLSTM)

s'1 s'k s'M. . .. . .

Content Interaction Layer

t1 tk tM. . .. . .

TP1 TP2 TP3 TP4 TP5

tp1

TP-sceneInteraction

Layer

. .

Content

. .

Salience wrtplotline

Final scenerepresentations

(b): Narrative structure prediction

tp2 tp3 tp4 tp5

(c): Summary scenes prediction

y1, ..., yk, ..., yM

(a): Scene encoding

. .

Figure 3: Overview of SUMMER. We use oneTP-specific attention mechanism per turning point inorder to acquire TP-specific distributions over scenes.We then compute the similarity between TPs and con-textualized scene representations. Finally, we performmax pooling over TP-specific similarity vectors andconcatenate the final similarity representation with thecontextualized scene representation.

are related to at least one of the key events inthe screenplay. We calculate the semantic similar-ity vi j of scene ti with TP t p j as in Equations (2)and (3). We then perform max pooling over vec-tors vi1, . . . ,viT , where T is the number of TPs(i.e., five) and calculate a final similarity vector v′ifor the ith scene.

The model is trained end-to-end on the summa-rization task using BCE, the binary cross-entropyloss function. We add an extra regularization termto this objective to encourage the TP-specific at-tention distributions to be orthogonal (since wewant each attention layer to attend to differentparts of the screenplay). We thus maximize theKullback-Leibler (KL) divergence DKL betweenall pairs of TP attention distributions t pi, i ∈ [1,5]:

O = ∑i∈[1,5]

∑j∈[1,5], j 6=i

log1

DKL(t pi

∥∥t p j)+ ε

(7)

Furthermore, we know from screenwriting theory(Hauge, 2017) that there are rules of thumb as to

when a TP should occur (e.g., the Opportunity oc-curs after the first 10% of a screenplay, Change ofPlans is approximately 25% in). It is reasonable todiscourage t p distributions to deviate drasticallyfrom these expected positions. Focal regulariza-tion F minimizes the KL divergence DKL betweeneach TP attention distribution t pi and its expectedposition distribution thi:

F = ∑i∈[1,5]

DKL (t pi‖thi) (8)

The final loss L is the weighted sum of all threecomponents, where a,b are fixed during training:L = BCE+aO+bF .

4 Experimental Setup

Crime Scene Investigation Dataset We per-formed experiments on an extension of the CSIdataset6 introduced by Frermann et al. (2018). Itconsists of 39 CSI episodes, each annotated withword-level labels denoting whether the perpetra-tor is mentioned in the utterances characters speak.We further collected scene-level binary labels in-dicating whether episode scenes are important andshould be included in a summary. Three humanjudges performed the annotation task after watch-ing the CSI episodes scene-by-scene. To facilitatethe annotation, judges were asked to indicate whythey thought a scene was important, citing the fol-lowing reasons: it revealed (i) the victim, (ii) thecause of death, (iii) an autopsy report, (iv) crucialevidence, (v) the perpetrator, and (vi) the motive orthe relation between perpetrator and victim. An-notators were free to select more than one or noneof the listed reasons where appropriate. We canthink of these reasons as high-level aspects a goodsummary should cover (for CSI and related crimeseries). Annotators were not given any informa-tion about TPs or narrative structure; the annota-tion was not guided by theoretical considerations,rather our aim was to produce useful CSI sum-maries. Table 2 presents the dataset statistics (seealso Appendix B for more detail).

Implementation Details In order to set the hy-perparameters of all proposed networks, we useda small development set of four episodes from theCSI dataset (see Appendix B for details). After ex-perimentation, we set the temperature τ of the soft-max layers for the TP-specific attentions (Equa-tion (5)) to 0.01. Since the binary labels in the

6https://github.com/EdinburghNLP/csi-corpus

Page 7: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

1926

overallepisodes 39scenes 1544summary scenes 454

per episodescenes 39.58 (6.52)crime-specific aspects 5.62 (0.24)summary scenes 11.64 (2.98)summary scenes (%) 29.75 (7.35)sentences 822.56 (936.23)tokens 13.27k (14.67k)

per episode scenesentences 20.78 (35.61)tokens 335.19 (547.61)tokens per sentence 16.13 (16.32)

Table 2: CSI dataset statistics; means and (std).

supervised setting are imbalanced, we apply classweights to the binary cross-entropy loss of the re-spective models. We weight each class by its in-verse frequency in the training set. Finally, in su-pervised SUMMER, where we also identify the nar-rative structure of the screenplays, we consider askey events per TP the scenes that correspond to anattention score higher than 0.05. More implemen-tation details can be found in Appendix C.

As shown in Table 2, the gold-standard sum-maries in our dataset have a compression rate ofapproximately 30%. During inference, we selectthe top M scenes as the summary, such that theycorrespond to 30% of the length of the episode.

5 Results and Analysis

Is Narrative Structure Helpful? We perform10-fold cross-validation and evaluate model per-formance in terms of F1 score. Table 3 sum-marizes the results of unsupervised models. Wepresent the following baselines: Lead 30% se-lects the first 30% of an episode as the summary,Last 30% selects the last 30%, and Mixed 30%,randomly selects 15% of the summary fromthe first 30% of an episode and 15% from thelast 30%. We also compare SUMMER againstTEXTRANK based on tf*idf (Mihalcea and Ta-rau, 2004), the directed neural variant describedin Section 3.1 without any TP information, avariant where TPs are approximated by their ex-pected position as postulated in screenwriting the-ory, and a variant that incorporates informationabout characters (Gorinski and Lapata, 2015) in-stead of narrative structure. For the character-based TEXTRANK, called SCENESUM, we substi-tute the fi, f j scores in Equation (6) with character-related importance scores ci similar to the defini-

Model F1Lead 30% 30.66Last 30% 39.85Mixed 30% 34.32TEXTRANK, undirected, tf*idf 32.11TEXTRANK, directed, neural 41.75TEXTRANK, directed, expected TP positions 41.05SCENESUM, directed, character-based weights 42.02SUMMER 44.70

Table 3: Unsupervised screenplay summarization.

F1 Coverageof aspects

# scenesper TP

Lead 30% 30.66 – –Last 30% 39.85 – –Mixed 30% 34.32 – –SUMMARUNNER* 48.56 – –SCENESUM 47.71 – –SUMMER, fixed one-hot TPs 46.92 63.11 1.00SUMMER, fixed distributions 47.64 67.01 1.05SUMMER, −P, −R 51.93 44.48 1.19SUMMER, −P, +R 49.98 51.96 1.14SUMMER, +P, −R 50.56 62.35 3.07SUMMER, +P, +R 52.00 70.25 1.20

Table 4: Supervised screenplay summarization; for inSUMMER variants, we also report the percentage of as-pect labels covered by latent TP predictions.

tion in Gorinski and Lapata (2015):

ci =∑c∈C [c ∈ S ∪ main(C)]

∑c∈C [c ∈ S](9)

where S is the set of all characters participating inscene si, C is the set of all characters participat-ing in the screenplay and main(C) are all the maincharacters of the screenplay. We retrieve the setof main characters from the IMDb page of the re-spective episode. We also note that human agree-ment for our task is 79.26 F1 score, as measuredon a small subset of the corpus.

As shown in Table 3, SUMMER achieves thebest performance (44.70 F1 score) among all mod-els and is superior to an equivalent model whichuses expected TP positions or a character-basedrepresentation. This indicates that the pre-trainednetwork provides better predictions for key eventsthan position and character heuristics, even thoughthere is a domain shift from Hollywood moviesin the TRIPOD corpus to episodes of a crimeseries in the CSI corpus. Moreover, we findthat the directed versions of TEXTRANK are bet-ter at identifying important scenes than the undi-rected version. We found that performance peakswith λ1 = 0.7 (see Equation (6)), indicating thathigher importance is given to scenes as the storyprogresses (see Appendix D for experiments withdifferent λ values).

Page 8: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

1927

In Table 4, we report results for supervisedmodels. Aside from the various baselines in thefirst block of the table, we compare the neuralextractive model SUMMARUNNER*7 (Nallapatiet al., 2017) presented in Section 3.2 with sev-eral variants of our model SUMMER. We exper-imented with randomly initializing the networkfor TP identification (−P) and with using a pre-trained network (+P). We also experimented withremoving the regularization terms, O and F (Equa-tions (7) and (8)) from the loss (−R). We as-sess the performance of SUMMER when we followa two-step approach where we first predict TPsvia the pre-trained network and then train a net-work on screenplay summarization based on fixedTP representations (fixed one-hot TPs), or alter-natively use expected TP position distributions aspostulated in screenwriting theory (fixed distribu-tions). Finally, we incorporate character-based in-formation into our baseline and create a supervisedversion of SCENESUM. We now utilize the charac-ter importance scores per scene (Equation (9)) asattention scores – instead of using a trainable at-tention mechanism – when computing the globalscreenplay representation d (Section 3.2).

Table 4 shows that all end-to-end SUMMER

variants outperform SUMMARUNNER*. Thebest result (52.00 F1 Score) is achieved by pre-trained SUMMER with regularization, outperform-ing SUMMARUNNER* by an absolute differenceof 3.44. The randomly initialized version withno regularization achieves similar performance(51.93 F1 score). For summarizing screenplays,explicitly encoding narrative structure seems tobe more beneficial than general representationsof scene importance. Finally, two-step versionsof SUMMER perform poorly, which indicates thatend-to-end training and fine-tuning of the TP iden-tification network on the target dataset is crucial.

What Does the Model Learn? Apart from per-formance on summarization, we would also like toexamine the quality of the TPs inferred by SUM-MER (supervised variant). Problematically, we donot have any gold-standard TP annotation in theCSI corpus. Nevertheless, we can implicitly assesswhether they are meaningful by measuring howwell they correlate with the reasons annotators citeto justify their decision to include a scene in thesummary (e.g., because it reveals cause of death

7Our adaptation of SUMMARUNNER that considers con-tent and salience vectors for scene selection.

or provides important evidence). Specifically, wecompute the extent to which these aspects overlapwith the TPs predicted by SUMMER as:

C=∑Ai∈A∑T Pj∈T P [dist(T Pj,Ai)≤1]

|A|(10)

where A is the set of all aspect scenes, |A| is thenumber of aspects, T P is the set of scenes inferredas TPs by the model, Ai and T Pj are the subsetsof scenes corresponding to the ith aspect and jth

TP, respectively, and dist(T Pj,Ai) is the minimumdistance between T Pj and Ai in number of scenes.

The proportion of aspects covered is given inTable 4, middle column. We find that coverage isrelatively low (44.48%) for the randomly initial-ized SUMMER with no regularization. There is aslight improvement of 7.48% when we force theTP-specific attention distributions to be orthogo-nal and close to expected positions. Pre-trainingand regularization provide a significant boost, in-creasing coverage to 70.25%, while pre-trainedSUMMER without regularization infers on aver-age more scenes representative of each TP. Thisshows that the orthogonal constraint also encour-ages sparse attention distributions for TPs.

Table 5 shows the degree of association be-tween individual TPs and summary aspects (seeAppendix D for illustrated examples). We observethat Opportunity and Change of Plans are mostlyassociated with information about the crime sceneand the victim, Climax is focused on the revelationof the motive, while information relating to causeof death, perpetrator, and evidence is captured byboth Point of no Return and Major Setback. Over-all, the generic Hollywood-inspired TP labels areadjusted to our genre and describe crime-relatedkey events, even though no aspect labels were pro-vided to our model during training.

Do Humans Like the Summaries? We alsoconducted a human evaluation experiment usingthe summaries created for 10 CSI episodes.8 Weproduced summaries based on the gold-standardannotations (Gold), SUMMARUNNER*, and thesupervised version of SUMMER. Since 30% ofan episode results in lengthy summaries (15 min-utes on average), we further increased the com-pression rate for this experiment by limiting eachsummary to six scenes. For the gold standard con-dition, we randomly selected exactly one scene

8https://github.com/ppapalampidi/SUMMER/tree/master/video_summaries

Page 9: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

1928

Turning Point Crime scene Victim Death Cause Perpetrator Evidence MotiveOpportunity 56.76 52.63 15.63 15.38 2.56 0.00Change of Plans 27.03 42.11 21.88 15.38 5.13 0.00Point of no Return 8.11 13.16 9.38 25.64 48.72 5.88Major Setback 0.00 0.00 6.25 10.25 48.72 35.29Climax 2.70 0.00 6.25 2.56 23.08 55.88

Table 5: Percentage of aspect labels covered per TP for SUMMER, +P, +R.

System Crime scene Victim Death Cause Perpetrator Evidence Motive Overall RankSUMMARUNNER* 85.71 93.88 75.51 81.63 59.18 38.78 72.45 2.18SUMMER 89.80 87.76 83.67 81.63 77.55 57.14 79.59 2.00Gold 89.80 91.84 71.43 83.67 65.31 57.14 76.53 1.82

Table 6: Human evaluation: percentage of yes answers by AMT workers regarding each aspect in a summary. Alldifferences in (average) Rank are significant (p < 0.05, using a χ2 test).

per aspect. For SUMMARUNNER* and SUMMER

we selected the top six predicted scenes basedon their posterior probabilities. We then createdvideo summaries by isolating and merging the se-lected scenes in the raw video.

We asked Amazon Mechanical Turk (AMT)workers to watch the video summaries for all sys-tems and rank them from most to least informa-tive. They were also presented with six questionsrelating to the aspects the summary was supposedto cover (e.g., Was the victim revealed in the sum-mary? Do you know who the perpetrator was?).They could answer Yes, No, or Unsure. Five work-ers evaluated each summary.

Table 6 shows the proportion of times partic-ipants responded Yes for each aspect across thethree systems. Although SUMMER does not im-prove over SUMMARUNNER* in identifying ba-sic information (i.e., about the victim and perpe-trator), it creates better summaries overall withmore diverse content (i.e., it more frequently in-cludes information about cause of death, evidence,and motive). This observation validates our as-sumption that identifying scenes that are semanti-cally close to the key events of a screenplay leadsto more complete and detailed summaries. Fi-nally, Table 6 also lists the average rank per system(lower is better), which shows that crowdwork-ers like gold summaries best, SUMMER is oftenranked second, followed by SUMMARUNNER* inthird place.

6 Conclusions

In this paper we argued that the underlying struc-ture of narratives is beneficial for long-form sum-marization. We adapted a scheme for identifyingnarrative structure (i.e., turning points) in Holly-wood movies and showed how this information

can be integrated with supervised and unsuper-vised extractive summarization algorithms. Ex-periments on the CSI corpus showed that thisscheme transfers well to a different genre (crimeinvestigation) and that utilizing narrative struc-ture boosts summarization performance, leadingto more complete and diverse summaries. Anal-ysis of model output further revealed that latentevents encapsulated by turning points correlatewith important aspects of a CSI summary.

Although currently our approach relies solelyon textual information, it would be interesting toincorporate additional modalities such as video oraudio. Audiovisual information could facilitatethe identification of key events and scenes. Be-sides narrative structure, we would also like to ex-amine the role of emotional arcs (Vonnegut, 1981;Reagan et al., 2016) in a screenplay. An often in-tegral part of a compelling story is the emotionalexperience that is evoked in the reader or viewer(e.g., somebody gets into trouble and then out of it,somebody finds something wonderful, loses it, andthen finds it again). Understanding emotional arcsmay be useful to revealing a story’s shape, high-lighting important scenes, and tracking how thestory develops for different characters over time.

Acknowledgments

We thank the anonymous reviewers for their feed-back. We gratefully acknowledge the support ofthe European Research Council (Lapata; award681760, “Translating Multiple Modalities intoText”) and of the Leverhulme Trust (Keller; awardIAF-2017-019).

ReferencesApoorv Agarwal, Sriramkumar Balasubramanian,

Jiehan Zheng, and Sarthak Dash. 2014. Parsing

Page 10: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

1929

Screenplays for Extracting Social Networks fromMovies. In Proceedings of the 3rd Workshop onComputational Linguistics for Literature, pages 50–58, Gothenburg, Sweden.

David Bamman, Brendan O’Connor, and Noah A.Smith. 2013. Learning latent personas of film char-acters. In Proceedings of the 51st Annual Meeting ofthe Association for Computational Linguistics (Vol-ume 1: Long Papers), pages 352–361, Sofia, Bul-garia.

David Bamman, Ted Underwood, and Noah A. Smith.2014. A Bayesian mixed effects model of literarycharacter. In Proceedings of the 52nd Annual Meet-ing of the Association for Computational Linguis-tics (Volume 1: Long Papers), pages 370–379, Bal-timore, Maryland.

John B Black and Robert Wilensky. 1979. Anevaluation of story grammars. Cognitive science,3(3):213–229.

Charles Oscar Brink. 2011. Horace on Poetry: The’ArsPoetica’. Cambridge University Press.

Daniel Cer, Yinfei Yang, Sheng-yi Kong, Nan Hua,Nicole Limtiaco, Rhomni St John, Noah Constant,Mario Guajardo-Cespedes, Steve Yuan, Chris Tar,et al. 2018. Universal sentence encoder. arXivpreprint arXiv:1803.11175.

Nathanael Chambers and Dan Jurafsky. 2009. Unsu-pervised learning of narrative schemas and their par-ticipants. In Proceedings of the Joint Conference ofthe 47th Annual Meeting of the ACL and the 4th In-ternational Joint Conference on Natural LanguageProcessing of the AFNLP, pages 602–610, Suntec,Singapore.

Jianpeng Cheng and Mirella Lapata. 2016. Neuralsummarization by extracting sentences and words.In Proceedings of the 54th Annual Meeting of theAssociation for Computational Linguistics (Volume1: Long Papers), pages 484–494.

James E Cutting. 2016. Narrative theory and the dy-namics of popular movies. Psychonomic bulletin &review, 23(6):1713–1743.

Yue Dong. 2018. A survey on neural network-basedsummarization methods. ArXiv, abs/1804.04589.

David K. Elson, Nicholas Dames, and Kathleen R.Mckeown. 2010. Extracting social networks fromliterary fiction. In Proceedings of the 48th AnnualMeeting of the Association for Computational Lin-guistics, pages 138–147, Uppsala, Sweden.

Alexander Fabbri, Irene Li, Tianwei She, Suyi Li, andDragomir Radev. 2019. Multi-news: A large-scalemulti-document summarization dataset and abstrac-tive hierarchical model. In Proceedings of the 57thAnnual Meeting of the Association for Computa-tional Linguistics, pages 1074–1084, Florence, Italy.Association for Computational Linguistics.

Syd Field. 2005. Screenplay: Foundations of Screen-writing. Dell Publishing Company.

Mark Alan Finlayson. 2012. Learning Narrative Struc-ture from Annotated Folktales. Ph.D. thesis, Mas-sachusetts Institute of Technology.

Lea Frermann, Shay B Cohen, and Mirella Lapata.2018. Whodunnit? crime drama as a case for naturallanguage understanding. Transactions of the Associ-ation of Computational Linguistics, 6:1–15.

Gustav Freytag. 1896. Freytag’s technique of thedrama: an exposition of dramatic composition andart. Scholarly Press.

Philip John Gorinski and Mirella Lapata. 2015. Moviescript summarization as graph-based scene extrac-tion. In Proceedings of the 2015 Conference ofthe North American Chapter of the Association forComputational Linguistics: Human Language Tech-nologies, pages 1066–1076, Denver, Colorado. As-sociation for Computational Linguistics.

Amit Goyal, Ellen Riloff, and Hal Daume III. 2010.Automatically producing plot unit representationsfor narrative text. In Proceedings of the 2010 Con-ference on Empirical Methods in Natural LanguageProcessing, pages 77–86, Cambridge, MA.

Max Grusky, Mor Naaman, and Yoav Artzi. 2018.NEWSROOM: A dataset of 1.3 million summarieswith diverse extractive strategies. In Proceedingsof the 16th Annual Conference of the North Amer-ican Chapter of the Association for ComputationalLinguistics: Human Language Technologies, pages708–719, New Orleans, USA.

Michael Hauge. 2017. Storytelling Made Easy: Per-suade and Transform Your Audiences, Buyers, andClients – Simply, Quickly, and Profitably. IndieBooks International.

Marti A Hearst. 1997. Texttiling: Segmenting text intomulti-paragraph subtopic passages. Computationallinguistics, 23(1):33–64.

Karl Moritz Hermann, Tomas Kocisky, EdwardGrefenstette, Lasse Espeholt, Will Kay, Mustafa Su-leyman, and Phil Blunsom. 2015. Teaching ma-chines to read and comprehend. In Advances in Neu-ral Information Processing Systems 28, pages 1693–1701. Morgan, Kaufmann.

Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. 2015.Distilling the knowledge in a neural network. arXivpreprint arXiv:1503.02531.

Anna Kazantseva and Stan Szpakowicz. 2010. Sum-marizing short stories. Computational Linguistics,36(1):71–109.

Joy Kim and Andres Monroy-Hernandez. 2015. Sto-ria: Summarizing social media content based onnarrative theory using crowdsourcing. CoRR,abs/1509.03026.

Page 11: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

1930

Diederik P Kingma and Jimmy Ba. 2014. Adam: Amethod for stochastic optimization. arXiv preprintarXiv:1412.6980.

Wendy G. Lehnert. 1981. Plot units and narrative sum-marization. Cognitive Science, 5(4):293–331.

Yang Liu and Mirella Lapata. 2019. Hierarchical trans-formers for multi-document summarization. In Pro-ceedings of the 57th Annual Meeting of the Asso-ciation for Computational Linguistics, pages 5070–5081, Florence, Italy. Association for Computa-tional Linguistics.

Interjeet Mani. 2012. Computational Modeling ofNarative. Synthesis Lectures on Human LanguageTechnologies. Morgan and Claypool Publishers.

Neil McIntyre and Mirella Lapata. 2010. Plot induc-tion and evolutionary search for story generation. InProceedings of the 48th Annual Meeting of the Asso-ciation for Computational Linguistics, pages 1562–1572, Uppsala, Sweden.

Rada Mihalcea and Hakan Ceylan. 2007. Explorationsin automatic book summarization. In Proceedingsof the 2007 joint conference on empirical methodsin natural language processing and computationalnatural language learning (EMNLP-CoNLL), pages380–389.

Rada Mihalcea and Paul Tarau. 2004. Textrank: Bring-ing order into text. In Proceedings of the 2004 con-ference on empirical methods in natural languageprocessing, pages 404–411.

Ramesh Nallapati, Feifei Zhai, and Bowen Zhou. 2017.Summarunner: A recurrent neural network based se-quence model for extractive summarization of docu-ments. In Thirty-First AAAI Conference on ArtificialIntelligence.

Ramesh Nallapati, Bowen Zhou, Caglar Gulcehre,Bing Xiang, et al. 2016. Abstractive text summa-rization using sequence-to-sequence rnns and be-yond. arXiv preprint arXiv:1602.06023.

Shashi Narayan, Shay B. Cohen, and Mirella Lapata.2018. Don’t give me the details, just the summary!topic-aware convolutional neural networks for ex-treme summarization. In Proceedings of the 2018Conference on Empirical Methods in Natural Lan-guage Processing, pages 1797–1807, Brussels, Bel-gium. Association for Computational Linguistics.

Sinno Jialin Pan and Qiang Yang. 2010. A survey ontransfer learning. IEEE Transactions on Knowledgeand Data Engineering, 22(10):1345–1359.

Pinelopi Papalampidi, Frank Keller, and Mirella La-pata. 2019. Movie plot analysis via turning pointidentification. In Proceedings of the 2019 Confer-ence on Empirical Methods in Natural LanguageProcessing and the 9th International Joint Confer-ence on Natural Language Processing (EMNLP-IJCNLP), pages 1707–1717.

Adam Paszke, Sam Gross, Soumith Chintala, Gre-gory Chanan, Edward Yang, Zachary DeVito, Zem-ing Lin, Alban Desmaison, Luca Antiga, and AdamLerer. 2017. Automatic differentiation in pytorch.

Patrice Pavis. 1998. Dictionary of the theatre: Terms,concepts, and analysis. University of Toronto Press.

Vladimir Iakovlevich Propp. 1968. Morphology of theFolktale. University of Texas.

Andrew J. Reagan, Lewis Mitchell, Dilan Kiley,Christopher M. Danforth, and Peter Sheridan Dodds.2016. The emotional arcs of stories are dominatedby six basic shapes. EPJ Data Science, 5(31):1–12.

Whitman Richards, Mark Alan Finlayson, andPatrick Henry Winston. 2009. Advancing com-putational models of narrative. Technical Report63:2009, MIT Computer Science and Atrificial In-telligence Laboratory.

David E. Rumelhart. 1980. On evaluating story gram-mars. Cognitive Science, 4(3):313–316.

Evan Sandhaus. 2008. The New York Times AnnotatedCorpus. Linguistic Data Consortium, Philadelphia,6(12).

Roger C. Schank and Robert P. Abelson. 1975. Scripts,plans, and knowledge. In Proceedings of the 4thInternational Joint Conference on Artificial Intelli-gence, pages 151–157, Tblisi, USSR.

Shashank Srivastava, Snigdha Chaturvedi, and TomMitchell. 2016. Inferring interpersonal relations innarrative summaries. In Proceedings of the 13thAAAI Conference on Artificial Intelligence, pages2807–2813, Phoenix, Arizona. AAAI Press.

Kristin Thompson. 1999. Storytelling in the new Holly-wood: Understanding classical narrative technique.Harvard University Press.

Tsvetomira Tsoneva, Mauro Barbieri, and Hans Weda.2007. Automated summarization of narrative videoon a semantic level. In International Conference onSemantic Computing (ICSC 2007), pages 169–176.IEEE.

Josep Valls-Vargas, J. Zhu, and Santiago Ontanon.2014. Toward automatic role identification in unan-notated folk tales. In Proceedings of the 10th AAAIConference on Artificial Intelligence and InteractiveDigital Entertainment, pages 188–194.

Paul Vicol, Makarand Tapaswi, Lluis Castrejon, andSanja Fidler. 2018. Moviegraphs: Towards under-standing human-centric situations from videos. InProceedings of the IEEE Conference on ComputerVision and Pattern Recognition, pages 8581–8590.

Christopher Vogler. 2007. Writer’s Journey: MythicStructure for Writers. Michael Wiese Productions.

Kurt Vonnegut. 1981. Palm Sunday. RosettaBooksLLC, New York.

Page 12: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

1931

Crime scene

12.4%Victim

14.6%

Perpetrator

14.4%

Cause of death

11.6%

Evidence36.1%

Motive

10.9%

Figure 4: Average composition of a CSI summarybased on different crime-related aspects.

Hao Zheng and Mirella Lapata. 2019. Sentence cen-trality revisited for unsupervised summarization.arXiv preprint arXiv:1906.03508.

Qingyu Zhou, Nan Yang, Furu Wei, Shaohan Huang,Ming Zhou, and Tiejun Zhao. 2018. Neural docu-ment summarization by jointly learning to score andselect sentences. arXiv preprint arXiv:1807.02305.

A Narrative Structure Theory

The initial formulation of narrative structure waspromoted by Aristotle, who defined the basictriangle-shaped plot structure, that has a beginning(protasis), middle (epitasis) and end (catastrophe)(Pavis, 1998). However, later theories argued thatthe structure of a play should be more complex(Brink, 2011) and hence, other schemes (Freytag,1896) were proposed with fine-grained stages andevents defining the progression of the plot. Theseevents are considered as the precursor of turningpoints, defined by Thompson (1999) and used inmodern variations of screenplay theory. Turningpoints are narrative moments from which the plotgoes in a different direction. By definition theseoccur at the junctions of acts.

Currently, there are myriad schemes describ-ing the narrative structure of films, which are of-ten used as a practical guide for screenwriters(Cutting, 2016). One variation of these modernschemes is adopted by Papalampidi et al. (2019),who focus on the definition of turning points anddemonstrate that such events indeed exist in filmsand can be automatically identified. Accordingto the adopted scheme (Hauge, 2017), there aresix stages (acts) in a film, namely the setup, thenew situation, progress, complications and higherstakes, the final push and the aftermath, separatedby the five turning points presented in Table 1.

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 11

32

34

36

38

40

42

44

F1 sc

ore

(%)

Directed neuralTextRankSUMMER

Figure 5: F1 score (%) for directed neural TEXT-RANK and SUMMER for unsupervised summarizationwith respect to different λ1 values. Higher λ1 valuescorrespond to higher importance in the next context forthe centrality computation of a current scene.

B CSI Corpus

As described in Section 4, we collected aspect-based summary labels for all episodes in the CSIcorpus. In Figure 4 we illustrate the average com-position of a summary based on the different as-pects seen in a crime investigation (e.g., crimescene, victim, cause of death, perpetrator, evi-dence). Most of these aspects are covered in10–15% of a summary, which corresponds to ap-proximately two scenes in the episode. Onlythe “Evidence” aspect occupies a larger propor-tion of the summary (36.1%) corresponding tofive scenes. However, there exist scenes whichcover multiple aspects (an as a result are anno-tated with more than one label) and episodes thatdo not include any scenes related to a specific as-pect (e.g., if the murder was a suicide, there is noperpetrator).

We should note that Frermann et al. (2018) dis-criminate between different cases presented in thesame episode in the original CSI dataset. Specif-ically, there are episodes in the dataset, where ex-cept for the primary crime investigation case, asecond one is presented occupying a significantlysmaller part of the episode. Although in the origi-nal dataset, there are annotations available indicat-ing which scenes refer to each case, we assume nosuch knowledge treating the screenplay as a singleunit — most TV series and movies contain sub-stories. We also hypothesize that the latent iden-tified TP events in SUMMER should relate to theprimary case.

Page 13: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

1932

Figure 6: Examples of inferred TPs alongside with gold-standard aspect-based summary labels in CSI episodes attest time. The TP events are identified in the latent space for the supervised version of SUMMER (+P, +R).

C Implementation Details

In all unsupervised versions of TEXTRANK andSUMMER we used a threshold h equal to 0.2 for re-moving weak edges from the corresponding fullyconnected screenplay graphs. For the supervisedversion of SUMMER, where we use additional reg-ularization terms in the loss function, we experi-mentally set the weights a and b for the differentterms to 0.15 and 0.1, respectively.

We used the Adam algorithm (Kingma and Ba,2014) for optimizing our networks. After experi-mentation, we chose an LSTM with 64 neurons forencoding the scenes in the screenplay and anotheridentical one for contextualizing them. For thecontext interaction layer, the window l for comput-ing the surrounding context of a screenplay scenewas set to 20% of the screenplay length as pro-posed in Papalampidi et al. (2019). Finally, wealso added a dropout of 0.2. For developing ourmodels we used PyTorch (Paszke et al., 2017).

D Additional Results

We illustrate in Figure 5 the performance (F1score) of the directed neural TEXTRANK andSUMMER models in the unsupervised setting withrespect to different λ1 values. Higher λ1 valuescorrespond to higher importance for the succeed-ing scenes and respectively lower importance for

the preceding ones, since λ1 and λ2 are bounded(λ1 +λ2 = 1).

We observe that performance increases whenhigher importance is attributed to screenplayscenes as the story moves on (λ1 > 0.5), whereasfor extreme cases (λ1 → 1), where only the laterparts of the story are considered, performancedrops. Overall, the same peak appears for bothTEXTRANK and SUMMER when λ1 ∈ [0.6,0.7],which means that slightly higher importance is at-tributed to the screenplay scenes that follow. In-tuitively, initial scenes of an episode tend to havehigh similarity with all other scenes in the screen-play, and on their own are not very informative(e.g., the crime, victim, and suspects are intro-duced but the perpetrator is not yet known). Asa result, the undirected version of TEXTRANK

tends to favor the first part of the story and the re-sulting summary consists mainly of initial scenes.By adding extra importance to later scenes, wealso encourage the selection of later events thatmight be surprising (and hence have lower simi-larity with other scenes) but more informative forthe summary. Moreover, in SUMMER, where theweights change in a systematic manner based onnarrative structure, we also observe that scenes ap-pearing later in the screenplay are selected moreoften for inclusion in the summary.

As described in detail in Section 3.3, we also

Page 14: Screenplay Summarization Using Latent Narrative Structure · 2020. 6. 20. · Screenplay Summarization Using Latent Narrative Structure Pinelopi Papalampidi1 Frank Keller1 Lea Frermann2

1933

infer the narrative structure of CSI episodes in thesupervised version of SUMMER via latent TP rep-resentations. During experimentation (see Sec-tion 5), we found that these TPs are highly corre-lated with different aspects of a CSI summary. InFigure 6 we visualize examples of identified TPson CSI episodes during test time alongside withgold-standard aspect-based summary annotations.Based on the examples, we empirically observethat different TPs tend to capture different typesof information helpful for summarizing crime in-vestigation stories (e.g., crime scene, victim, per-petrator, motive).