Extracting semantic properties to ... -...

26
i University of Amsterdam FNWI Artificial Intelligence Bachelorthesis 24-07-2012 Extracting semantic properties to capture the dynamic structure within the motifs of the Thematic Model Student: Rosemary Novita Moerbeek, 5942756 Supervisor: Frank Nack

Transcript of Extracting semantic properties to ... -...

Page 1: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

i

University of Amsterdam FNWI

Artificial Intelligence

Bachelorthesis

24-07-2012

Extracting semantic properties to capture the

dynamic structure within the motifs of the Thematic

Model

Student: Rosemary Novita Moerbeek, 5942756

Supervisor: Frank Nack

Page 2: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

1

Page 3: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

2

Index

1 Abstract 3

2 Introduction 3

2.1 Narratives, Themes and Motifs 3

2.2 Structure and Culture 5

3 Related Work 6

4 Method 7

4.1 Function 7

4.2 Weight 8

4.2.1 Derivation 8

4.3 Combination 10

4.3.1 Derivation 10

4.4 Position 11

4.4.1 Derivation 11

4.5 Order 12

4.5.1 Derivation 12

4.6 Commotion 13

4.6.1 Derivation 13

4.7 Prototyping 17

5 Evaluation 17

5.1 First Narrative 17

5.2 Second Narrative 20

6 Conclusion 23

7 References 24

Page 4: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

3

1 Abstract This thesis aims to tackle the problem of capturing semantic structures within narratives,

using a distinguishable approach bridging the semantic gap. Using the components of

Hargood’s Thematic Model [1], we try to capture the patterns of features denoting motifs

using five semantic properties (weight, combination, position, order, commotion) which can

be applied to a pattern matching algorithm to automatically generate narrative structures.

2 Introduction

Stories are part of our everyday live. We are naturally communicating and receiving

experiences from direct and indirect audiences and sources through a variety of media.

However, it appears difficult to establish an objective, effective and machine readable model

of narratives. This thesis presents a method that can improve a recently found machine

readable model of narratives with the purpose of partly eliminating subjectivity from the

original model. To understand the original model as well as the improved version, the next

sections will enlighten some background motivations.

2.1 Narratives, Themes and Motifs

The terms stories and narratives are sometimes used interchangeable, however, they are not

exactly the same. Structuralist theories define a narrative as composed of any series of human

experiences [2]. Furthermore, narratives can be deconstructed into a story and a discourse [3]

where the story represents a chronology of all the information to be communicated and the

discourse defines what parts of the story are told and how those arts are presented.

Narratives are a way of presenting information in its broadest sense, whether to inform,

persuade or entertain. They are used to make sense of information and experiences to

ourselves and to others and they apply to every sort of medium.

Narrative systems are systems that inhabit functionalities that disseminate, create or analyze

narratives. Since narratives play a great role in our everyday life, narrative systems that help

us process these narratives more efficiently would be very valuable. Let’s consider a few

future possibilities.

1) Information could be presented more engaging to the user by adjusting the structure –

not the content- to his or her personal preferences.

2) A semantic structure could be assigned to the unstructured results of a search engine

query, helping the user find his way through the data overload we encounter every

day.

3) Computer games could adapt and regenerate their plot based on the actions of the

player.

The idea of a narrative system is not new. Systems exist that can analyze, represent or

generate narratives. However, the main limitation of these systems is that their generated

Page 5: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

4

narratives do not have the cohesion of a human-made narrative [4]. A plot is said to be

coherent when a human interpreter can understand the way in which the events of a narrative

have meaning and relevance to the outcome of the story [5].

A large amount of research is done on the literal content of the narrative such as the plot [5] or

the used language [6]. However, little attention has been given to the notion of themes in

narratives despite its contribution to the richness and cohesiveness of narratives. Themes

allow the author to transmit a certain context without having to explicitly mention the

components of that context. The reader already knows the components and only needs the

theme as a trigger to link them to the narrative. Most narrative systems have no proper

concept of what a theme is, how it can be represented or interpreted.

Hargood [1] recently tried to cover this section of narratology by setting up the Thematic

Model which is a representation model for themes using semiotic term expansion. The choice

for semiotic term expansion suits themesbest due to the implicit nature of a theme. Increasing

the understanding of such a theme might lead to narrative systems that create more coherent

and rich narratives.

The general assumption of the model is that a theme, being a conceptual entity, is build up

from a collection of other themes or motifs which, in context of each other, connote the

theme. Motifs are directly denoted by features, which are the visible computable elements

(either by authored tags, metadata or

computer vision detected annotations).

The use of the terms connotation and

denotation may seem subtle, but turns

out to be essential for the model. Note

that the idea of connotation is that signs

have meaning beyond their literal

expression and therefor the denotative

sign becomes a signified [7]. For

example, one may connote the concept

of a celebration from the sign of a

wedding cake, rather than just denoting

the concept of a cake. The ability of

making such a connotation arises from

our cultural and social background and

from our experiences, which makes a

connotation a subjective notion.

Fig. 1: The Thematic Model[1]

The Thematic Model is not capable of representing stronger and weaker relations between

features and motifs and motifs and themes, although this evidently applies in realistic

situations. Furthermore, the model cannot represent certain combinations of signs denoting

Page 6: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

5

the signified. Serious problems could arise from this limitation. For example, one might want

to represent the motif diner as denoted by the combined features spoon, fork and knife. With

this representation, it is not enough to detect just a knife in the natom. An additional fork and

spoon has to be detected as well to denote the motif diner.

Another limitation arises when the model is used to represent motifs and themes over a series

of natoms which together form a narrative. It is not possible to represent the dynamics of this

narrative. One might want to represent a motif over a series of natoms as denoted by a certain

pattern of features rather than just the presence of these features. The order, quantity or

position of features is not expressible, even though these contain valuable information which

could lead to a more accurate denotation.

In this thesis, a method will be presented that dynamically captures the semantic pattern of a

narrative by assigning five semantic properties to the features to which a pattern recognition

algorithm can be applied. Section 4 shows how this function is derived. The general idea is

that, with these properties, we can produce a narrative from a new dataset simulating the

pattern of a known narrative or known genre, while updating and adjusting the function if

needed for the next narrative generation.

Let’s take a closer look at how this method works. Assume we start with an approved, linear

structured narrative, consisting of a number of natoms. The natoms we’ll use are photographs,

but the method works uniformly for other sorts of media. For all features that we can detect in

the natoms, a separate function is created. After analyzing the narrative, all created functions

together represent the semantic pattern of that narrative and since it was assumed to be an

approved narrative, this pattern is usable for narrative generation. However, we will not use

the abstract properties of the pattern as input for a supervised learning algorithm. Instead, we

will introduce five semantic properties, all derived from the semantic pattern mentioned

above. This choice is based on the fact that for a learning system to succeed, the

representation must be crafted with knowledge about the application domain.

2.2 Structure and Culture

The composition of narratives is subject to cultural differences in (at least) two ways:

intentional structure differences and attentional state differences [8]. Intentional structure

refers to the relations between actions and consequences within the narrative. The extent to

which these relations are comprehensible determines the perceived cohesion of the story.

However, there are great cultural differences in what people find normal reactions, depending

on their personal experiences and knowledge from their environment. For example, someone

from the western culture may find it perfectly understandable that a homeless dog gets

adopted by a caring family, while this situation would not be so obvious for someone from a

culture that do not keep animals as pets.

The attentional state is an abstraction of the focus of attention as the narrative unfolds,

modeled by a set of focus spaces. Since people from different cultures are used to different

attentional states in stories, they often have problems sympathizing with subjects of the story

Page 7: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

6

when this known pattern is changed. It is no coincidence that all western fairytales start with

an introduction of their main characters (setup) before a dramatic problem arises (conflict)

which has to be defeated (resolution) to live happily ever after.

The method presented in this thesis has no difficulties with the above mentioned problems

because the functions adapt to the user’s habits and preferences.

3 Related work

Much work is done on modeling a narrative structure. One of the first works on structuralist

concepts is done by Propp [9], who identified 8 character archetypes and a sequence of 31

functions characterizing folk tales. The deconstructing of narratives in their component

elements expresses in his work was one of the first contributions to automatic story

generation.

For fully automatic narrative generation to succeed, the task of automatic photo annotation

must first be accomplished. This is important for the training phase as well as the actual

generation. The paper of Tuffield presents the Photocopain system [10], which is a semi-

automatic photo annotation system which combines the metadata of the photograph with a

source of information from various information extracting tools. Combining such a photo

annotator which the method presented in this thesis, would highly expand the pool of usable

natoms which in turn can be used to refine the prototype patterns.

Using the modeled structure of narratives, Screenplay System’s software program Dramatica

actually generates narratives, guided by concrete directions of the ‘human writer’. The writer

makes the choices, and Dramatica matches it with a known narrative structure from its

enormous database. Choices have to be made until just one structure matches the directions.

However, this kind of knowledge based system becomes very fragile when faced with a

dynamic problem domain or any task it was not specifically designed to do [11].

The work of Brooks [12] presents a computational narrative model which divides the task into

three functional areas: “defining an abstract narrative structure, defining a collection and

organization of story pieces with some representation of their meaning and a navigational

strategy or reasoning through that collection of story pieces”. It defines an elegant framework

including all parties involved in automatic narrating: the artist, the agent, the audience and the

story itself. Furthermore, it identifies three environments mapping to the three functional

areas mentioned earlier. These environments articulate important boundaries that are

implicitly used in this thesis. The method presented in this thesis distances itself from the

environment referred as Structural Environment by Brooks. This is the environment in which

the structure of the narrative is described in simple abstract terms. In this thesis, we reside in

the other two environments of Brooks: the Representational Environment, in which

knowledge of the various story elements is captured in the form of relationships between story

elements and the Presentational Environment, in which software agents sequence the different

story elements according to an agent’s individual stylistic preferences.

Page 8: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

7

4 Method

The following sections will present a method that dynamically captures the semantic pattern

of a narrative by assigning five semantic properties to the features. The general idea is that,

with these properties, we can produce a narrative from a new dataset simulating the pattern of

a known narrative or known genre, while updating and adjusting the function if needed for the

next narrative generation.

The method is based on the machine learning assumption that a machine is able to predict

target labels with a sufficient set of features. The quality of feature selection is therefore of

great influence on the results of the predictor.

For the sake of consistency, all components of the method are listed below together with their

description as used in this particular context [1].

Narrative: a structured, purposeful, communication of an experience to an audience.

Natom: is a `narrative atom'. A singular irreducible component of storytelling, where further

reduction is either impossible or would cause the natom to no longer make sense. E.g. a

photograph, a sentence or short paragraph, a short video clip.

Feature: is identifiable evidence of the presence of a particular piece of content within a

narrative or piece of narrative. This might be a piece of metadata, a tag, or automatically

extracted keywords or descriptions.

Motif: is an atomic component of a theme. Motifs are generalizations or classifications of

elements with the narrative that connote a theme, they are directly denoted by features within

the narrative. E.g. Snow, Rose, Grave, Champagne.

Theme: is a high level concept that is part of the subtext of a narrative. A theme is not directly

present within the narrative but is connoted through the presence of motifs. E.g. Winter, Love,

Death, Celebration.

Property: is a meta-feature; a feature of a feature.

4.1 Function

To be able to use the data extracted from the narrative (a series of natoms), a function is fitted

onto the data points of every single tagged feature. A semantic pattern is expresses by the

combinations of all function together. However, to apply a simple pattern recognizing

algorithm on the separate functions is not likely to produce usable results. The features must

not be seen as isolated units since the relations between them contain the information we are

seeking. But even a pattern recognizer that takes several functions as input will not uncover a

workable semantic pattern. This problem arises from the concept of a semantic gap which is

the abstraction of the difference between two descriptions of the same object by different

representations. Two different representations can never represent the exact same semantic

meaning. Therefore, if we want to extract semantics that are meaningful to the human mind,

we need to bridge the semantic gap going from a computational representation of the narrative

Page 9: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

8

to a human understandable representation. This can only be done by using properties that

already capture some semantics as an input for a machine learning algorithm. Five of these

semantic properties are listed below. Note that they represent features in a machine learning

context, but to avoid confusion with the higher level features of the thematic model, we will

call them properties.

4.2 Weight

An important property of a feature is how strong it relates to a motif. That is, with what level

of certainty does the feature connote a motif? One can easily imagine that the feature snow

fall connotes the motif snow with a very high certainty, while the feature bitter sky has a much

lower certainty of connoting the same motif (feature examples used from Hargood [1]). This

may seem trivial, but assigning weight factors to features makes it possible to assign a level of

confidence to the connotation of the whole motif (usually connoted by more than one feature).

Consequently, it becomes possible to keep track on various alternatives and assign definite

connotations after the whole narrative is analyzed. Combined with a simple path finding

algorithm, this metadata could avoid making certain connotation mistakes in analyzing tasks,

which otherwise would have resulted in noise. In creation tasks however, this property is even

more important. It gives the user the opportunity to define a certain threshold with which he

or she can set the level of obviousness of the created narrative. A high threshold produces a

more cliché story, while a lower threshold causes the production of a story that is somewhat

more difficult, but perhaps more interesting to read.

4.2.1 Derivation

Calculating weight factors is done for every motif separately. The input is defined by the

features denoting M. If these features are already known, the derivation works optimal. If they

are not exactly known on forehand, one can just insert a sequence of features of which the

denoting features are members (but not known). In the latter case, a very low weight will be

given to the features not denoting the motif, and a much higher weight will be given to the

sequence of features denoting the motif. A threshold can be applied to discover this sequence.

Page 10: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

9

Calculating weight factors for the features of motif M

Input:

Sequence of features denoting motif M: [f1, f2, …, fq]

w(f1), .. w(fq) 1/q % set initial weights

alpha range [0.1, 0.001] % learning rate

n 1 % natoms counting

for j 1 to q % for all features

xj integral(fj) % count appearances

yj avg(xj) % get average appearances

if(yj<xj) % if fj appears more than average

zj xj-yj % get difference

w(fj) w(fj) + alpha*zj % adjust weight (increase)

for k 1 to q % for all other features

if(k /= j) % not the same as fj

w(fk) w(fk)–(alpha*zj)/(q-1) % adjust weight (decrease)

end for % end for

else % if fj appears less than average

zj yj-xj % get difference

w(fj) w(fj) – alpha*zj % adjust weight (decrease)

for(h 1 tot q) % for all other features

if(k /= j) % not the same as fj

w(fh) w(fh)+(alpha*zj)/(q-1) % adjust weight (increase)

end for % end for

avg(xj) (xj/n)+avg(xj)

n = n + 1

end for

Page 11: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

10

4.3 Combinations

Some features cannot connote a motif by themselves; they need another feature, or any

number of features, to be present in the same natom to connote a motif. This includes also the

possibility that a natom needs to appear twice or more in the same natom. It is easy to find

examples of this phenomenon in real life. It takes more than just a spoon to denote a diner and

one swallow does not make a summer.

Therefore, the combination property is introduced. It captures the dependences between all

features, but needs only be assigned to features which have any dependency relations. For

every dependency relation, a weight is given between 0 and 1.

4.3.1 Derivation

We now have a vector for every feature which specifies for every feature (including the same

feature) how often they appear in the same natom on a continuous scale from 0 to 1. It is

optional to set thresholds to define when a combination is mandatory (for example, v[q] >

0.8) or just common (0.3 < v[q] > 0.8).

Input:

Sequence of all features: [f1, f2, …, fq]

sequence of all natoms: [n1, n2, .. nm]

for j 1 to q

vj null vector with length q

for k 1 to m

if fj is in nm

For s 1 to q

if fs is in nm % if the two features combine in this

natom

v[s] v[s] + count(fs) % put number of appearances in vector

else

v[s] v[s] % features do not combine

End for

End for

normalize vj so that all entries sum up to 1

End for

Page 12: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

11

4.4 Position

The position in which features appear is of great importance to the semantics of the story. An

object appearing in the beginning of a narrative is more likely to present a main character than

something that gets introduced just before the narrative ends. This assumption at least holds

for classical structured stories [12] and stories following the Freytag’s pyramid structure [13],

where the beginning of the story is used to define the setting for the rest of the story. After

this introduction (also called exposition or setup), a problem arises (conflict or rising action).

The most important factors are introduced in these two first acts. The position property is

introduced to capture this semantic information.

Its contribution to creation tasks is more evident than to analyzing tasks. However, even in

analyzing tasks this property does reduce the probability of making an incorrect motif

classification. In creation tasks, this property is very important. It ensures that features which

are important to the narrative are indeed introduced in the first or second act. This property is

thus responsible for a human understandable arrangement of the created narrative.

4.4.1 Derivation

The durations of the separate acts are not known. This means no sharp line can be drawn

between ‘important objects introduced in the beginning of the narrative’ and ‘less important

objects introduced just before the end of the narrative’.

Therefore, we will use a function with a smooth transition from 1 to 0:

Fig 2: Position function

The position property is calculated by applying the position function to the first occurrence of

every feature. It is important to only consider the first occurrence, since main characters and

objects are likely to appear throughout the whole narrative, which would obviously reduce

their position property and that is just what we do not want here.

An example:

A certain object is introduced in the third natom of a sequence of 30 natoms.

Page 13: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

12

Another object is introduced in the twentieth natom of the same sequence:

4.5 Order

What does it mean for a feature to be preceded by another feature? Sometimes nothing of

course, but sometimes a lot. The order property captures the structure within the motif. It can

be the case that a certain feature is only (partly) connoting a motif, if it is preceded by another

(series of) feature(s). This property is most similar to the combination property, but there is

one essential difference: the order property takes time into account. One could argue that the

combination property is only a special case of the order property: namely, the case that the

time between the appearances of two (or more) features, is negligible. The reason to express

them nonetheless as separated properties is that they have very different semantic meaning.

With the order property, we can express both the notion of causality and as well as

conditionality.

4.5.1 Derivation

Input:

Sequence of all features: [f1, f2, …, fq]

sequence of all natoms: [n1, n2, .. nm]

k window size

for j 1 to q

vj null vector with length q

zj null vector with length q

for k 1 to m

if fj is in nm

For s 1 to q

if fs appears in the last k natoms

v[s] v[s] + count(fs) % put number of appearances in vector

else

v[s] v[s] % features do not combine

End for

End for

normalize vj so that all entries sum up to 1

End for

Page 14: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

13

4.6 Commotion

Some features cause more commotion than others. This property is introduced to distinguish

between background objects and foreground objects. While a background object (e.g. a tree,

chair or sky) can be important to a motif, it usually is not important to the semantics of the

narrative. A background object sometimes appears often in a sequence of natoms, but is only

part of the context of the narrative and does not add meaning to the story being told.

Misjudging such a background object is not difficult: a tree can appear frequently in a

narrative which plays in the woods, acquiring high weighting properties, but has nothing to do

with the meaning of the narrative.

The commotion property assigns a high value (ranging from 0 to 1) to features that cause

some commotion in the narrative; that is, if it causes a significant change in the semantic

pattern.

4.6.1 Derivation

To be able to detect a sudden change in the semantic pattern, a similarity measure is needed.

Consider the following data fragment:

Feature Natom1 Natom2 Natom3 Natom4 Natom5

1 1 1 1 0 0

2 0 0 3 0 0

3 2 1 1 4 5

4 1 0 0 6 3

5 0 1 0 2 3

Figure 3: Semantic pattern

A sudden change in the pattern can be noticed at natom4, while the first three natoms are

much alike. In semantic terms, natom4 shows a reaction to an earlier action. The calculation of

the commotion property can be divided in two separated tasks: detecting a significant change

in the semantic pattern and finding features that could be responsible for this change to

happen. Below, we will discuss each task apart.

Page 15: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

14

4.6.1.1 Detecting change

In default, a window of one natom is used to compare with the next window. It is optional to

increase both windows, for example to increase robustness of the method. It is convenient to

apply bigger windows for films (a few seconds) or long texts (one or two paragraphs), but for

narratives with relatively short sequences of natoms, we will use one natom as window size.

A cosine similarity function is applied to express similarity between natoms.

function detect_commotion % detects significant changes between natoms

% in vector form

input:

sequence of natoms [n1, n2, .. nm]

q = number of features

t = threshold for similarity [-1 to 1]

for i 1 to m

get feature vector vi % entries of vector from semantic pattern

get feature vector vi+1 % next natom

if sim > t % no significant change

proceed with next natom

else

get_commotion_features % execute second task

Page 16: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

15

4.6.1.2 Detecting responsible feature(s)

function get_commotion_features

input:

natomi showing significant change

sequence F of available features [f1,f2, .. fq]

sequence C of initial commotion property [c1, c2,.. cq]

n number of natoms to check % default: 4

t threshold for required slope

set all property values to 0

for m 1 to q

for j 1 to n % for every natom to check

v vector with length n

v[j] f(i-j) % number of appearances in preceding natom

end for

% Next, we calculate the slope of the data points (least square)

if slope > t % check to filter out unrelated features

cm cm + 1 % increase commotion property of fm

end for

normalize C

end

end for

Page 17: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

16

4. 7 Prototyping

A valuable application of the properties as mentioned above is the specification of prototypes

or genres in terms of the properties. Having a prototype of a certain story type (e.g. drama,

documentary, detective), captured with the five introduced properties, will definitely help

generating similar structured stories, without losing the ability to create new structures or

contents as is the case with knowledge based narrative generators. If sufficient labeled data

can be analyzed and recorded, it becomes possible to create a narrative of a genre by matching

the prototype pattern with new data. An elegant collaboration of narratology and machine

learning has more potential in accomplishing this task than both academic fields would have

alone.

5 Evaluation

The operating of the method presented in this thesis will be showed using short sequences of

natoms that nonetheless have narrative structure and meaning. Analyzing larger sequences

works similarly and may result in even more accurate findings, but to avoid losing grip on the

semantics, smaller sequences are preferred for evaluation matters. Consider the following

narratives, each consisting of four natoms:

(1a) (1b) (1c) (1d)

(2a) (2b) (2c) (2d)

Page 18: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

17

5.1 First narrative

The first sequence portrays some common daily rituals from sunrise until evening. One starts

the day with a hot cup of coffee, does some cleaning activity, spends some time behind a pc

and ends the day with cooking. Captured by the thematic model would look like the following

scheme.

Fig 4.: Data applied to model

After the annotation, listed in the above table, the properties can be extracted, following the

directions as specified in chapter 4.

Feature N1a N1b N1c

N1d

Hand

Cup

Coffee

Sponge

Glove

Foam

Plate

Keyboard

Computer

Knife

Cuttingboard

Onion

1

1

1

0

0

0

0

0

0

0

0

0

1

0

0

1

1

1

1

0

0

0

0

0

1

0

0

0

0

0

0

1

1

0

0

0

1

0

0

0

0

0

0

0

0

1

1

1

Page 19: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

18

5.1.1 Weight property

The following table lists the adjusted weights of each feature after analyzing the whole

narrative with learning rate 0,1

Fig 5.: assigning weighted values to features

The most interesting part of the result is the hand feature. This feature appears in every single

motif, suggesting its appearance is independent from each of the four motifs. The results are

showing the same trend: the weight of the hand feature is decreasing, while the weights of all

other denoting features are increasing. Furthermore, when a lot of different features are

denoting the same motif, they get a lower initial weight assigned. In this evaluation, only one

example per motif is analyzed, but after analyzing a larger number of motif examples, this

drawback is neutralized.

Feature MotifCoffee Motifcleaning MotifPC MotifCooking

Hand

Cup

Coffee

Sponge

Glove

Foam

Plate

Keyboard

Computer

Knife

Cuttingboard

Onion

0.333

0.408

0.408

0

0

0

0

0

0

0

0

0

0.0125

0

0

0.275

0.275

0.275

0.275

0

0

0

0

0

0.333

0

0

0

0

0

0

0.408

0.408

0

0

0

0.0625

0

0

0

0

0

0

0

0

0.325

0.325

0.325

Page 20: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

19

5.1.2 Combinations property

A representative selection of the combination properties are expressed in the scheme below.

Note that it is a relational property and that the relations are not necessarily symmetric.

Fig 6.: Schematic drawing of combinational relations

5.1.3 Position property

The position property is meaningless for this particular sequence of natoms, because it thrives

on the assumption that the narrative has the classic dramatic construction (setup, conflict,

resolution). This assumption does not apply on this story, so we will not consider the position

property here.

Page 21: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

20

5.1.4 Order property

The order property is expresses in a relation between features. When two features often

precede each other, their relation gets a higher weight. In this short example, all relations still

have the same weight, because they appear only once. The duration between their appearances

is recorded as well. The scheme below represents the order relations between its features.

Fig 7.: Schematic drawing of ordering relations

5.1.5 Commotion property

The commotion property is, for other reasons than the position property, also not appropriate

to apply on this set of natoms. The commotion property needs smoother transitions between

the natoms to be meaningful. A close look at the semantic pattern reveals that all four

columns diverge exactly the same from each other. This means that after each natom, the

same portion of commotion arises.

Page 22: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

21

5.2 Second narrative

The second narrative portrays a typical fishing day: getting onto a boat, doing some angling,

catching a fish and eating the caught fish.

Fig 8.: semantic pattern of the second narrative

Fig 9.: Second narrative modeled

5.2.1 Weight property

Weight properties are meaningful for almost every sequence of natoms. Deriving the weight

properties for the second narrative gives the following results.

Fig 10.: Assigning weights to the features of second narrative

Feature N2a N2b N2c N2d

Boat

Water

Fishing rod

Fish

Baked fish

Plate

1

1

0

0

0

0

1

1

3

0

0

0

0

1

1

1

0

0

0

0

0

0

1

1

Feature Motifboating Motifangling Motifcatch Motifbakedfish

Boat

Water

Fishing rod

Fish

Baked fish

Plate

0.550

0.525

0

0

0

0

0

0.525

0.700

0

0

0

0

0.358

0.333

0.408

0

0

0

0

0

0

0.575

0.575

Page 23: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

22

5.2.2 Combination property

The non-symmetric property of the relations expresses a natural understanding: if there is

water, it is not very often the case that there is also a fish, but if we see a fish, it is probably

swimming in the water. The combination relations are captured in the following scheme.

Fig 11.: Schematic drawing of the combinatorial relations

5.2.3 Commotion property

Looking at figure 8, one can easily see that the last column is very different from the others.

The method for finding commotion properties detects such a sudden change in the pattern

using a similarity measure to compare frames. The next step is to seek responsible features for

this sudden change. This is done by calculating the slope of the fitted function onto the

semantic pattern of every feature that precedes the last natom. With this approach, the features

boat and water drop off and the method finds that the feature fish or fishing rod was

responsible for the commotion, which is true.

Page 24: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

23

6 Conclusion

Automatic storytelling is an ambitious but tempting task. With possible applications in every

media related domain, its relevance is beyond doubt. The method presented in this thesis aims

to contribute towards this ultimate goal, distinguishing itself from related work by linking the

essences of narratology and machine learning.

The property evaluating appears promising as the results correspond to our own knowledge

and beliefs about narratives. However, it is yet too early for a definite conclusion. Due to the

subjectivity of the correctness of a generated narrative, conclusions cannot be drawn without

considering the subjects.

In this thesis, we approach the Semantic Gap by putting semantics in the properties that can

be used as inputs for a machine learning algorithm. The property evaluation suggests that this

is a fruitful approach since it follows human intuition without either using static knowledge

based rules or a large amount of abstract properties.

A disadvantage of the method which became known in the evaluation is the fact that not all

properties are equally meaningful for all narratives. This is the case with abstract properties as

well, but because of our small set of properties, this disadvantage is more serious here. It

forces developers to select training data with great care, which can be time consuming.

As this method is situated between the academic fields of narratology and machine learning, it

operates like an interface between the two. Integration from both sides will take it to the next

level.

Page 25: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

24

7 References

[1] Hargood (2011), “Semiotic Term Expansion as the Basis for Thematic Models in

Narrative Systems”, University of Southampton, PhD Thesis

[2] M. McQuillan. The Narrative Reader. Routledge, London, 2000

[3] R. Barthes and L. Duisit. An introduction to the structural analysis of narrative. New

Literary History, 1966.

[4] T. Murray, T. Shen, J. Piemonte, C. Condit, and J. Thibedeau. Adaptivity for conceptual

and narrative flow in hyperbooks: The metalinks system. Adaptive Hypermedia and Adaptive

Web-Based Systems, 2000.

[5] M. Riedl and R. Young. Character-Focused Narrative Generation for Execution. Virtual

Storytelling, Vol. 2897Springer Berlin / Heidelberg ,2003.

[6] M. Cavazza and F. Charles. Dialogue generation in character-based interactive

storytelling. In AAAI First Annual Articial Intelligence and Interactive Digital Entertainment

Conference, California, USA, ,2005.

[7] R. Barthes. Mythologies. Editions du Seuil, 1957.

[8] B. Grosz , and C. L. Sidner in Comput. Linguist., Vol. 12 Cambridge, MA, USA: MIT

Press, p. 175—204, 1986.

[9] Propp. Morphology of the folktale. Austin, University of Texas Press, 1928

[10] M.Tuffield, S. Harris, D.P. Dupplaw, A. Chakravarthy, C. Brewster, N. Gibbins, K.

O'Hara, F. Ciravegna, D. Sleeman, N.R. Shadbolt, and Y. Wilks. Image Annotation with

Photocopain, in 'Proceedings of the First International Workshop on Semantic Web

Annotations for Multimedia (SWAMM)’, 2006.

[11] Screenplay-Systems, Dramatica. Screenplay Systems: Burbank, California, 1994.

[12] K. M. Brooks, Do Story Agents Use Rocking Chairs? The Theory and Implementation of

One Model for Computational Narrative., in Philippe Aigrain; Wendy Hall; Thomas D. C.

Little & V. Michael Bove Jr., ed., 'ACM Multimedia' , ACM Press, , pp. 317-328 , 1996.

[13] Freytag, G. (1965 (Nachdruck; EA 1863)), Die Technik des Dramas , Darmstadt .

Page 26: Extracting semantic properties to ... - staff.science.uva.nlstaff.science.uva.nl/~bredeweg/pdf/BSc/20112012/Moerbeek.pdf · be applied to a pattern matching algorithm to automatically

25