INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account...

176
INFORMATION TO USERS This manuscript has been reproduced frorn the microfilm master. UMI films the text directly from the original or copy submitted. Thus, some thesis and dissertation copies are in typewriter face, while others may be from any type cf cornputer printer. The quality of this reproduction is dependent upon the quality of the copy submitted. Broken or indistinct print, colored or poor quality illustrations and photographs, print bleedthrough, substandard margins, and improper alignment can adversely affect reproduction. In the unlikely event that the author did not send UMI a complete manuscript and there are missing pages, these will be noted. Also, if unauthorized copyright material had to be removed, a note will indicate the deletion. Oversize materials (e-g., rnaps, drawings, charts) are reproduoed by sedioning the original, beginning at the upper left-hand corner and wntinuing from left to right in equai sections with small overlaps. Each original is aiso photographed in one exposure and is inciuded in reduced form at the back of the book. Photographs inciuded in the original manuscript have been reproduced xerographically in this copy. Higher quality 6" x 9" black and white photographic pnnts are available for any photographs or illustrations appearing in this copy for an additional charge. Contact UMI directly to order. Bell & Howell Information and Leaming 300 North Zeeb Road, Ann A W r , MI 48106-1346 USA 800-521 -0600

Transcript of INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account...

Page 1: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

INFORMATION TO USERS

This manuscript has been reproduced frorn the microfilm master. UMI films the

text directly from the original or copy submitted. Thus, some thesis and

dissertation copies are in typewriter face, while others may be from any type cf

cornputer printer.

The quality of this reproduction is dependent upon the quality of the copy

submitted. Broken or indistinct print, colored or poor quality illustrations and

photographs, print bleedthrough, substandard margins, and improper alignment

can adversely affect reproduction.

In the unlikely event that the author did not send UMI a complete manuscript and

there are missing pages, these will be noted. Also, if unauthorized copyright

material had to be removed, a note will indicate the deletion.

Oversize materials (e-g., rnaps, drawings, charts) are reproduœd by sedioning

the original, beginning at the upper left-hand corner and wntinuing from left to

right in equai sections with small overlaps. Each original is aiso photographed in

one exposure and is inciuded in reduced form at the back of the book.

Photographs inciuded in the original manuscript have been reproduced

xerographically in this copy. Higher quality 6" x 9" black and white photographic

pnnts are available for any photographs or illustrations appearing in this copy for

an additional charge. Contact UMI directly to order.

Bell & Howell Information and Leaming 300 North Zeeb Road, Ann A W r , MI 48106-1346 USA

800-521 -0600

Page 2: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input
Page 3: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

NOTE TO USERS

The original manuscript received by UMI contains pages with indistinct andlor slanted print. Pages were microfilmed as

received. r

This reproduction is the best copy available

Page 4: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input
Page 5: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Designed Intelligence: A Language Teacher Model

1. Staatsexamen, Weingarten University, 1983

II. Staatsexamen, Weingarten University, 1985

M a s e r of Arts, Simon Fraser University, 1993

THE REQUIREMENTS FOR THE DEGREE OF

DOCTOR OF PHILOSOPHY

in the Department

Linguistics

O Gertrud Doris M- He* 1998

S M O N FRASER UNIVERSITY

July 1998

All rights reserved. This work may not be

reproduced in whole o r in part, by photocopy

or other means, without permission of the author.

Page 6: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

National Library j*I of Canada Bibliothèque nationale du Canada

Acquisitions and Acquisitions et Bibliographie Services services bibliographiques

395 Wellington Street 395, rue Wellington Ottawa ON K I A ON4 OttawaON K1AON4 Canada Canada

Your Iiie Voire relerence

Our nle Notre rëfërence

The author has granted a non- L'auteur a accordé une licence non exclusive licence dowing the exclusive permettant à la National Library of Canada to Bibliothèque nationale du Canada de reproduce, loan, distribute or seIl reproduire, prêter, distribuer ou copies of this thesis in microfom, vendre des copies de cette thèse sous paper or electronic formats. la forme de microfiche/film, de

reproduction sur papier ou sur format électronique.

The author retains ownership of the L'auteur conserve la propriété du copyright in this thesis. Neither the droit d'auteur qui protège cette thèse. thesis nor substantial extracts fkom it Ni la thèse ni des extraits substantiels may be printed or otherwise de celle-ci ne doivent être imprimés reproduced without the author's ou autrement reproduits sans son permission. autorisation.

Page 7: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Abstract

This thesis presents the design of an Intelligent Language Tutoring

System that models a language teacher by evaluating and responding to

student performance o n foreign language exercises. The design is

implemented in Head-driven Phrase Structure Grammar and illustrated with

German.

A review bf the literature provides an ovenriew of Intelligent

Language Tutors, examining existing systems with respect to error coverage,

user scope, and computational generality. Design criteria are reevaluated,

emphasizing a pedagogically-infomed, student-centered approach.

The design presented facilitates this view, d s o solving several extant

problems in Intelligent Language Tutoring Systems: errors need not be

anticipated. Additionally, the techniques successfully address ambiguous

readings, contingent and ambiguous errors. Finally, the system is modular

making it adaptable to different languages, specinc student needs, andlor to

reflect the pedagogy of a particular language instructor.

The design of the Intelligent Language Tutoring System consists of - five components: The Domain Knowledge, the Licensing Module, the Analysis

Module, the Student Model, and the Filtering Module. The Domain

Knowledge provides the linguistic analysis, the remaining components add

pedagogical value.

The Domain Knowledge consists of a parser with a grammar which

parses sentences and phrases t o produce sets of phrase descriptors. A phrase

descriptor is a mode1 of a particular grammatical phenornenon. It records

whether o r not the grammatical phenornenon is present in the input and Kso,

whether it is correctly or incorrectly formed. The Licensing Module selects one

of the possible parses by taking into account the likelihood o f the error. The

..- Designed Intelligence: A hnguage Teacher Mode1 111

Page 8: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Analysis Module detides on the source of the error. It takes a phrase

descriptor as input and generates sets of possible responses to the leamer's

input that the instruction system can use when interacting with the student.

The Student Model records mastery of grammatical structures as well as

structures with which the leamer has problems. The level of the learner

according to the current state of the Student Model, determines the particular

feedback displayed. Finally, the Filtering Module decides o n the order of the

instructional feedback by considering the importance of an error in a given

exercise and the dependency between syntacticdly higher and lower

constituents.

- -

Designed Intelligence: A Language Teacher Model iv

Page 9: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Acknowledgments

Joy and despair visited me in completing this work, but my constant

companions were my supervisor, my f d y ~ and my friends. You always had

another nice word to say to keep me going.

Paul, you devoted much of your time and knowledge to my success;

such insights as can be found in this work are due to your outstanding efforts

and guidance as my supervisor. You were a wonderfiil person t o work with.

Many, many th&.

Thanks, too, to my cornmittee members, especially Fred Popowich, for

the thoughtfid commentary.

To ali my fkiends and fellow graduate students, all of whom 1 have not

seen much lately, thanks for your support and patience. And to the monitors in

the Language Leaming Centre, thanks for being so considerate during the

time when my foms was on this work.

To my f d y in Germany, my heartfelt thanks for a l l your care and

thoughts for me.

And Chris and Bodhi, during the little time I had to spare you gave

me aIi the support, love, and energy needed to complete this work. Our life

suffered, but k d y my mind can be with you again. LYTD.

Designed Intelligence: ALanguage Teacher Mode1 v

Page 10: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Table of Contents

. . Approvd ........................................................................................................... 11

... Abstract ........................................................................................................... iii

Acknowledgements ........................................................................................ v

List of Figures ...fi ........................................................................................ ix

CHAPTER 1 Intelligent Language Tutoring Systems . 1

1.1 Introduction ........................................................................ 1

........................... 1.1.1 Natural Language Processing 4

.................................................. 1.1.1.1 Syntactic Parsers 5

1.1.2 ParskgIll-Formedhputin

........... Intelligent Language Tutoring Systems 7

............... 1.1.2.1 Anticipating Ill-Formed Student Input 8

.................................... 1.2 Augmented Transition Networks 12

1.3 Intelligent Language Tutorhg Systems Using

1.4 Evaluation of Intelligent Language Tutoring Systems .. 28

........................................................... 1.5 The German Tutor 30

... ......................... 1.5.1 The Domain Knowledge .. 32

1.5 -2 The Licensing Module ............................ .. ..... 37

1.5.3 The halys is Module ......................... .... ..... 37

.......................................... 1.5.4 The Seiident Mode1 40

....................................... 1.5.5 The Filtering Module 41

Designed Intelligence: A Language Teacher Mode1 vi

Page 11: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

CHAPTER 2

CHAPTER 3

............................... Analyzing Student Input 45

Introduction ...................................................................... 45

Errors in Feature Matchîng ............................................. 50

General Technique for Errors in

Feature Matching ............................................. 50

............................................ Ambiguous Errors 55

Two Likely Sources of an Error ......................... 56

......................... An Unlikely Source of an Emor 57

Analyzing Noun Phrases ................................. 60

Feature Percolation ................... .... ............. 62

........... Phrase Descriptors and Feature Values 67

............................................ Errors in Linear Precedence 69

2.3.1 General Technique for Verb-Second Position .. 70

................... 2.3.2 Verb-Second Position .,. ................ 72

.... 2.3 -3 General Technique for Verb-Final Position 75

.......................................... 2.3.4 Verb-Fin al. Position 77

........................................................................ Summary 80

Disambiguating Multiple Sentence Readings .................................................................. 83

Introduction .................................................................... 83

Licensing Module ........................................................ 88

3.2.1 Subject-Head and Head-Subject Licensing .... 90

3.2.2 Head-Subject-Complement and

................................. Head-Subject Licensing 93

3.2.3 Licensing vs . Fewest Error Parse .................... 96

......................................................... Ad Hoc Techniques 99

. ............................... 3.3.1 Finite Verbs vs hibitives 99

Designed Intelligence: A Language Teacher Mode1

Page 12: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

3.3.2 Summary ....................................................... 105

.......... CHAPTER 4 Pedagogically-Informed Feedback 107

Introduction

...................................................... 4.2 The Analysis Module 113

......................................................... 4.3 The Student Mode1 119

................................ 413.1 Implicit Student Models 121

................................ 4.3.2 Explicit Student Models 122

4.3.3 Feedback Suited t o Leamer Expertise ......... 124

..................................................... 4.4 The Filtering Module 127

.............................. 4.4.1 The Error Priority Queue 130

.......................................... 4.4.2 Contingent Errors 132

4.5 Summary ....................................................................... 138

CHAPTER 5 Conclusion ....................................................... 140

5.1 Summary ........................................................................ 140

.................................... 5.2 Further Research .... . . 144

App endix

References

Designed Intelligence: A Language Teacher Mudel viii

Page 13: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

List of Figures

Figure 1.1 : An Augmented Transition Network ............................................ 13

Figure 1.2 : The German Tutor .............................. .. ...................-................ 31

Figure 1.3 : Partial Lexical Entry for geht ................. .... ............................ 33

Figure 1.4 : M a r h g Number Features for Singular Nouns ....................... .. 35

Figure 1.5 : M a r h g Number Features for Plural Nouns ............................. 35

Figure 1.6 : Recordhg Number Features forgeht ....................................... 36

Figure 1.7 : DATR Code Listing for a Finite Verb in a Main Clause ............ 39

Figure 2.1 : Partial Feature Structure for geht .................... .. .................... 46

Figure 2.2 : Lexical Entry for g e b t .................... .. ........................................ 51

Figure 2.3 : Lexical Entry for er ................................................................. 51

Figure 2.4 : Marking Person Features for er ........................ .. .................. 52

Figure 2.5 : M a r b g Person Features for gehst .................... ....... ........ 53

Figure 2.6 : Marking Person Features for du .................... .. ...................... 53 ......... Figure 2.7 : Case, Number, and Gender Features ........................ .. 60

............................... Figure 2.8 : Lexical Entry for der ..... .......................... 62 Figure 2.9 : Lexical Entry for Gotter .......................................................... 64

.......................... Figure 2.10 : Lexical Entry for der Gotter ........................ .. 65

Figure 2.11 : Lexical Entry for zürnen ....................... .. .............................. 66 Figure 2.12 : Phrase Descriptors for Der Gotter ziinzen ................................. 66

Figure 2.13 : Partial Structure of Finite Verb Position .................................. 71

Figure 2.14 : Feature Values for Verb-Second Position ................................. 71

............... Figure 2.15 : Percolation Pattern for Er geht ............................. .. 72

Figure 2.16 : Unary Assigngos Rule ............................................................. 73 ............................. Figure 2.17 : Phrase Descriptor for Verb-Second Position 73

............... ................... Figure 2.18 : Percolation Pattern for Sie ihm hi&! .. 75

-

Demgned Intelligence: A Language Teacher Mode1 ix

Page 14: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Figure 2.19 : Partial Structure of Nonfinite Verb Position ............................ 76

Figure 2.20 : Feature Values for Verb-Final Position ................................... 76

Figure 2.21 : Percolation Pattern for Er kann gehen .................................... 77

.............................................................. Figure 2.22 : Unary Assigngos Rule 78

Figure 2.23 : Phrase Descriptor for Verb-Final Position ................................ 78

Figure 2.24 : Percolation Pattern for Sie Rann Mfen KZaus ......................... 79

Figure 3.1 : Two Analyses for Er fragt das Maddzen ..................................... 84

Figure 3.2 : Two Analyses for Er sieht die Frau ............................................. 91

........................................................................ Figure 3.3 : Subject-Head Rule 92

Figure 3.4 : Two Analyses for *Heute sieht sie er ......................................... 94

............................................................. Figure 3.5 : Lexical En* for g e k n 101

............................................................... Figure 3.6 : Lexical Entry for kann 103 .

Figure 3.7 : Phrase Descriptors forgehen as a Finite Verb ......................... 104

Figure 4.1 : Pedagogic Modules ..................................................................... 108

Figure 4.2 : Granularity Hierarchy for Constraints in Feature Matching . . 115

Figure 4.3 : DATR Code Listing for a Finite Verb in a Subordkate Clause 116

Figure 4.4 : Granularity Hierarchy for Constraints in Linear Precedence . 118

................................... ................................ Figure 4.5 : Contingent Errors .... 135

Remgned Intelligence: A hgriage Teacher Mode1 s

Page 15: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

CHAPTER I Intelligent Language Tutoring Systems

1.1 Introduction

Cornputer-Assisted Language Learning (CALL) is the result of the

convergence of several fields of research addressing the use of computers in

language processing. Whde the influence of computational linguistics and

machine translation may be indirect, modern CALL s y stems draw heavily

upon the kdings of these two areas. The handling of naturd language by the

computer contributes greatly to the fluency of interaction between the human

user and the machine.

We k d a direct influence on CALL in experiments in Programmed

Instruction (PI). In the late sixties, CALL systems were primarily developed

on large-scale, maidkame syçtems in universities where computer sessions

were intended to replace classroom instruction. Programmed Instruction .. -. - .

Designed Intelligence: A Language Teacher Mode1 1

Page 16: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

proposed the view that the best way of learning a task is to split it into small

units, where the successfd completion of one building-block ieads to the next.

The generally sound pedagogical principle of dividing a large Iearning

tasir into conceptudy smaller units was, however, distorted by an over-

emphasis on rote mernorization manifested in repetitive drills, multiple choice

answers, and uninformative feedback. For example, in a typical drill exercise,

the student was a&ed to type in the answer which was checked as each

character was entered. If any error resulted, the computer assumed control

and typed out the word "wrong" Barker & Yeates 19851.

CALL questioned the effectiveness of such systems. A program with a

simple letter t o letter match is incapable of dserentiating types of errors: not

only is it, therefore, incapable of providing any valuable, evaluative feedback,

but, in ignoring the source of the error when selecting another problem, it

relies upon an inflexible, program-centered, rather than student-centered,

definition of clBiculS. m e , in considering the early systems, one should

make allowances for the limitations of the then current technology, the fad

remains that, although PI has been refhed over the years, it has never

achieved a high degree of popularity P r i c e 19911. The original h e a r

programs (representing fixed sequences of instruction) improved, becoming

more sophisticated branching programs; most, however, remained based on

multiple choice answers. Inasmuch as the new is oRen an offshoot of the old,

as among teaching approaches which emulate some procedures of the previous

approach while rejecting others, CALL is hiçtorically related to PI, adopting

however a concern for individualized learnuig, self-pacing, and immediate

feedback.

Due to an increased emphasis of carrent teaching approaches on the

student as an active participant in the learning process, increasingly more

Designed Intelligence: A Language Teacher Mode1 2

Page 17: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

attention is beîng paid to the interactive aspects of CALL systems Burns,

Parlett, and RecEeld 19911. As a consequence, CALL research has shifted its

focus from drill-and-practice to tutoring systems. I n tutoring programs, the

computer is still the "judge-of-the-right-ansmer? But as opposed to drill-and-

practice applications, the path to the right answer involves a fair amount of

student choice, control, and in particular, student-compter interaction

m s c h a u e r 19961.

Marshall Cl9881 identifies the significant interactive palities of

Computer-Assisted Language Learning as one advantage of irnplementing the

computer into the language classro om. l'rue interaction, however, requires

intelligent behaviour on the part of the computer. Without intelligence, the

system is merely another method of presenting information, one not especially

preferable t o a static medium like print. Instead of multiple choice questions,

relatively uninformative answer keys, and gross mainstreaming of students

characteristic of workbooks, modern C U L is aixning a t interactive computer

systems possessing a high degree of &cial intelligence and capable of

processing Naturd Language input molland et al. 19931. For a program to be

intelligent, however, i t must emulate the way a Ianguage teacher evaluates a

student response.

A language teacher fkst examines the sentence and locates the error.

Once the error has been identified, s/he decides on the reason for the error. In

some instances, the source of an error might not be easily determined due t o

error ambiguity. For this reason, a teacherys evaluation of the error will take

into account fùrther factors. These might include the student's previous

performance history &or the Gequency and difnculty of a particular

grammatical construction. The final result of this process is error-contingent

feedbackl suited to learner expertise.

Designed Intelligence: A Language Teacher Mode1 3

Page 18: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

The goal of this dissertation is to design a mode1 of a language

instructor by evaluating and responding to student performance o n foreign

language exercises. The steps taken by a language instructor in the correction

process are achieved by the computational analysis described and

implemented, by way of example, in the Geman Tutor, an Intelligent

Language Tutoring System (ILTS) for German.

- The &st step ... in the evaluation process of students' input is the

sentence analysis. In ILTSs, this is handled by the parser and the grammar.

The parser analyzes a sentence according to the knowledge of the language

encoded in the grammar. In section 1.1.1, I will discuss the importance of

Natural Language Processing (NLP) to ILTSs. In section 1.1.2,I will introduce

existing techniques for parsing ill-forrned student input, a challenging task

not unique to ILTSs.

1.1.1 Natural Language Processing

The strength of NLP is that it dows for a sophisticated error analysis

where student tasks . can go beyond multiple-choice questions andor f2.l-in-

the-blanks. Simple drills are based on str ing matching algorithms, that is, the

student response is compared letter for letter against an answer key. However,

one obviously cannot enter the arbitrarily many sentences required for

meaningfd practice into memory for purposes of cornparison. NLP provides

the analytical complexiG underpinning an ILTS.

The pedagogicd goal behind an ILTS is t o provide error-contingent

feedback suited to learner expertise. For example, if a student chooses an

incorrect article in Gerrnan the error might be due t o incorrect infiection for

1. "F'eedback tailored to the nature of the student's error is called error-contingent feedback." Alessi & Trollip 119851, p. 116.

Designed Intelligence: A Language Teacher Mode1

Page 19: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

gender, number, or case. In such an instance the program must be capable of

distinguishing between the three error types. For cognitive leaniing to occur,

instructional feedback must address the different sources of an error

mumehart & Norman 1975, Venezky & Osin 19911.

The error analysis performed by the computer system f o m s the basis

for error-contingent feedback. Garrett Cl9871 describes four kinds of feedback:

1. "presents only the correct answer,

2. pinpoints the location of errors on the basis of the cornputer's letter-by ietter cornparison of the student's input with the machine-stored correct version (pattern markup),

3. based on analysis of the anticipated wrong answers, error messages associated with possible errors are stored in the computer and are presented if the studentrs response matches those possible enors (error-anticipation technique),

4. most sophisticated, uses an intelligent, Natural Language Processing (NLP) approach such as the "parsing" technique in which the computer does a linguistic analysis of the student's response,

Studies [Nagata 1991, 1995, 1996, van der Linden 19931 addressing

the question of what End of feedback a computer program should give have

shown that not only do students appreciate the more sophisticated feedback

made possible by NLP, but also perform better on the language skills being

taught. The fact that students learn better provides the rationale for

emploping parsers in Computer-Assisted Language Learning.

1.1.1.1 Syntactic Parsers

Natural Lanmage parsers show great promise in the syntactic

analysis of students' input. They are best employed for ILTçs in introductoryl

intermediate language courses where the focuç is primady on form rather

than on content. At the advanced level where a stronger emphasis is placed on

content, syntactic parsers become less usefiil because they are less reliable.

2. Garrett [1987], as cited in Nagata 119911, p. 330,

Designed Intelligence: A Language Teacher Mode1 5

Page 20: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Inevitably, grammatical constructions become more elaborate and ambip~ous

at the higher levels which causes difficulties for parsers in general.3

A rider of studies have proven the usefuhess of syntactic parsers

in language learning. For example, a study conducted by Juozulynas Cl9941 at

Miami University showed that only 20% of errors in the essays of second-year

students of German are of a semantic nature.4 Juozulynas collected 349

students compositions. In d, 360 pages (313 essays) were included in the

study. The error distribution in his study was:

syntax: 28.6%

morphology: 24.4%

punctuation: 12.3%

spelling: 14.7%

semantics: 20%

A study by Rogers Cl9841 who collected 26 Gerrnan essays with an

average length of 569 words revealed similar results. Her distribution of

errors is:

syntax: 35%

morphology: 34.5%

lexical: 15.6%

orthography: 9.5%

complete transfer of EngIish expression: 5.4%

Adjusting Rogers' error classification to match ~ u o ~ y n a s ~ , 30% of

errors are of semantic origin. The higher percentage of semantic errors in

Rogers' study might be due to the fact that "the Miami University student

samples were Gom secondyear students, while the students in Rogers' study

3. There are a number of ILTSs which focus on communicative language leaming and thus emphasize fluency over accuracy. See Holland et al. [1995al and Swartz & Yazdani Cl.9921. 4. Juozulynas classified semantic errors as errors of meaning, such as wrong word choice, "made-upn words, and errors in pronoun reference. 5. For an even cornparison, Juozu lpas took Rogers' classification and assigned her categones lexical errors, complete transfer of EngZish expression, and some types of syntactic and morphological enors to the semantic category.

Designed Intelligence: A Language Teacher Mode1 6

Page 21: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

were advanced, with at least f o u years of learning Geman in a formal

environment, in many cases supplemented by visits to ~ e r m a n ~ . " ~

Given the outcornes of these studies, syntactic parsers can treat a

large percentage of student errors and thus are a powerfd tool in second

language learning.7 The following section, section 1.1.2, will discuss the

techniques employed by existing ILTSs in handling dl-fonned input.

. .

1.1.2 Parsing Ill-Formed Input in Intelligent Language Tutoring Systems

Although parsing ill-fomed input has been a challenge for all NLP

applications, ILTSs diEer fundamentally fkom other NLP applications in how

and why they handle ill-formed input. In most applications, the goal is to

successfidly parse and analyze a sentence, despite any errors. In ETSs,

however, the focus lies on tracing student's language knowledge rather than

o n the linguistic analysis of well-formed input. Such systems are inherently

prescriptive; for error-contingent feedback, they need to analyze where the

student deviated from the expert knowledge. In this respect, LTSs have a

more difficult task. Fortunatelx however, they usually do not deal with the

relatively large input domain found in other NLP applications.

In ILTSs, errors occur, not because the studentys howledge is a strict

subset of the expert knowledge, but because the learner possesses knowledge

potentially different in quantity and quality. For other NLP applications,

Weischedel & Sondheimer 119831 analyze errors as either absolute o r relative.

Absolute errors refer to errors in the language output. Examples are tense

and number mistakes, o r word order problems. Relative errors address

6. Juozulynas f19941, p. 16. 7. See also Sanders C1991J.

Deçigned Intelligence: A Langaage Teacher Mode1 7

Page 22: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

grammatical structures which are correct but are beyond the scope of the

grammar.

Although this definition of absolute errors addresses the concems of

an ILTS, there is nonetheless a Meren t goal behind the two conceptions:

ILTSs aim at providing the student with error-contingent feedback. Thus

ILTSs need to analyze student errors, and not simply withstand them. The

error -analysis performed by an ILTS provides the source of an error and thus

enables the student to leam grammatid constructions of the target

language. Simply withstanding errors fails the purpose of a paner-based

KTS.

Relative errors refer to language the user possesses that is beyond the

system's knowledge, which is rarely the case with an Intelligent Language

Tutor. ILTSs focus on a language subset as described in students' grammar

books. It is a diagnostic tool for language learners to improve their second

language grammar skills. The system responds t o a student's work on the

basis of any errors the student may have made and the model of the student

which the system has constructed to that point.

1.1.2.1 Anticipating Ill-Formed S tudent Input

ILTSs augment grammars that parse grammatical input in one of

three ways. Each method overcomes some obstacles; however, ail have

concomitant disadvantages which will be discussed in the context o f specific

implernentations in section 1.2. The following is a brief description of the

three general approaches.

In the h t technique, ILTSs may augment d e s such that if a

particular rule does not succeed, specific error routines, that is meta-iules8,

force application of the ni le by systernatically relaxhg its constraints. Meta-

Designed Intelligence: A Language Teacher Mode1 8

Page 23: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

rules are commody employed for agreement errors. An example of a meta-

rule as d e h e d by Chen and Kurtz Cl9891 is given in (cl:

(la) sentence(s(Np, Vp, Num)) --> noun-phrase(Np, Num),

verb-phrasewp, Num).

Cb) verb-phrase(vp(vN), Num) --> intrans-verbW, Num).

(c) intrans-verbw, Num, Relax-flag) -->

it-verbCV, Num); . .

Relax-flag == true, it-verbW, WrongNum),

WrongNum == Nurn,

error-fiaggingCX, subject-verb-agreement,

TV, Num, ~ r o n ~ ~ u m l ) . ~

The g r m a r rules given in (la) and Cb) specZy that a sentence

consists of a subject and an intransitive verb which agree in number. If the

noun and the verb do not agree in number, the parse would simply fail. The

meta-rule given in (cl, however, relaxes the eonstraint on number and will

allow the pane t o proceed. The predicate errorfluggiing will add an error

message and the student can be informed that a rnistake in number has

occurred.

In a second approach, ILTSs may augment the grammar with d e s

which are capable of parsing ill-formed input (buggy rules)1° and which apply

if the grammatical iules fail. Buggy rules are distinct fiom meta-rules in that

they do not force application of the same d e , but rather, provide a distinct

rule for every ill-formed construction. For example, in German the paçt

participle always occurs sentence-finally in a main-clause, given in example

(2):

8. The term meta-nile used in this dissertation does not refer to schema for other context-free d e s as used i n grammar formalisms such as GPSG- 9. The meta-de given has been adapted from Chen & Kurtz 11989:58-91 who provide a d e for a transitive verb. 10. Buggy d e s are sometimes called mal-rules.

Designed Intelligence: A Language Teacher Mode1 9

Page 24: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

( 2 ) Der Mann hat das Buch &. The man has read the book.

Due to native language interference'', English leamers of German

will tend to place the past participle in the position which reçembles English

word order, that is, between the auxiliary and the noun phrase. A buggy d e

would anticipate precisely this enor by describing a structure where the past

participle incorrectly occurs in the position derived fkom the English n o m ,

succe~sfüLly parsai the ungrammatical sentence and providing error-

contingent feedback.

Finallx with feature grammar formalisms, lLTSs may also alter the

unification algorithm itself such that in the event of contlicting feature values

the parse does not fail, but instead applies a different set of procedures.

Parsers designed for language instruction typically contain

components which anticipate o r search for errors in the event that the

grammatical rules are not successfd, buggy rules being a common instance.

Searching for errors, however, presents a number of problems.

First, due to the vast error scope and unpredictability of some errors,

ill-formed input can only be partially anticipated [Yazdani & Uren 19881. The

fewer errors anticipated, the srnaller the error coverage. Sentences containing

errors which have not been anticipated by the designer cannot be processed by

the system, inevitably resulting in generic rather than error-contingent

feedback, and failùlg to achieve one of the primary goals of an ILTS.

11. Native language interference, or interlingual transfer, refers to the use of elements fiom one Ianguage while speakinglwriting another [Richards 19711. In leaming a second language, the Iearner maps onto previously existing cognitive systems, which usually is the native language. The school of Contrastiue Analysis [Lado 19571 t i e d to e~pla in errors through reference to the native language of the learner. However, this view was rejected by later work and studies in Errorhalysis [Corder 1967, Selinker 19721. See also Larsen-Freeman & Long C19911.

Designed Intelligence: A Language Teacher Mode1 10

Page 25: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Second, anticipating errors lacks generalit~ Most systems base the

search for errors on the native language of the student. Any system which

anticipates errors requires that the same error in the target language is

presented in each different source language. The fewer source languages

considered the more user-limited the system. For example, a system which

anticipates errors made by English learners of German, will tend not t o

handie errors speciiically - . made by French leamers of German.

For the reasons of error coverage, generality, and user scope, the

methods employed in handling incorrect student - input in ILTSs should be

those that provide the most general treatment of errors, and that make

minimal use of emor anticipation.

The methods available t o detect errors are determined by the

programming language and the grammar formalism. KTSs have been

implemented in both procedural and declarative programrning languages-

Programming language to some extent determines the choice of grammar

formalism, which in turn determines what detection methods c a n be applied,

and how errors are perceived in the system Matthews 1993,1994, Sanders &

Sanders 19891.

The following @scussion wiU focus on LTSs which have been

implemented in either Augmented Transition Networks Woods 1970 1 or

declarative representations.12 The systems will be discussed +th respect to

the method used to detect errors and their e r r o r scope.13

12- A m , in their original design, were conceived as parsing rather than grammar formalisms. However, they have developed into a grammar-l3~e concept, especially in computational linguistics [Shieber 19861. 13. The systems discussed al1 place a pedagogical focus on form rather than content.

Designed Intelligence: A Language Teacher Mode1 11

Page 26: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

1.2 Augmented Transition Networks

Transition Networks (TNs) consist of a set of states, movement

between which is controlled by rules according t o the next elernent in the

input. States are referred to as nodes; transitions between states are referred

t o a s arcs. TNs are ipherently procedural.

Augmented Transition Networks (ATNs) additiondy employ

registers which hold grammatical information. ATNs also allow actions to be

associated with each arc, for instance, the setting of a register, o r the caUing

of another network.

The Augmented Transition Network Uustrated in Figure 1.1 parses

sentences by following the arcs, and accepting, from the input string,

constituents that are on the arc labels. The ATN can either accept a word, a

proper noun for example, o r it can c d another entire subnetwork, VP. Parsing

proceeds und a node is reached in which parsing can stop, indicated by the

double circle. In addition, the arcs set and test register NUM to control

number agreement between the subject and the verb.

The ATN given in Figure 1.1 parses sentences such as Mary sleeps by

going through the following steps:

Start in state 1 of S network. Accept NP (Mary, NUM = ssingular). Set register NUM of NP = singular. Set register NUM of S = singular.

Proceed to state 2 of S network, Cal1 VP network. Start in state 1 of VP network, Accept V (sleeps, NUM = singular).

Set register NUM of VP = singular.

Designed Intelligence: A Language Teacher Mode1 12

Page 27: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

NP VP set NUM of S test NUM of S =NUMofNP =NUMofVP

S 7 t n n

Figure 1.1: An Augmented Transition Network [Covington 19947

Proceed to state 2 of VP network and exit.

Test that NUM of S = NUM of VP

(the test succeeds since both have the value singular).

Proceed to state 3 of S network and exit.

While t h e Augmented Transition Network will successfully parse

Mary sleeps, note, however, that the sentence Mary sleep will fail since the

register on number agreement is violated. Violations of this End, however, are

common among second language students. An ILTS needs to parse such

sentences, and analyze the errors t o provide error-contingent feedback.

Meta- and buggy rules are the most common procedures employed for

handling students' errors in ILTSs. In ATN systems, agreement constraints

are commonly relaxed with meta-des, while buggy-des are used for errors

in word order. Buggy d e s might also make up a complete second g r m a r ,

the student's native language. In this case, meta-rules are used for analyzing

overgeneralization errors14.

The pioneer work in ETSs was developed by Weischedel et al. [1978].

The system is a prototype German tutor implemented as an Augmented

Designed Intelligence: A Language Tacher Mode1 13

Page 28: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Transition Network with a semantic and syntactic component. The authors

address four classes of errors:

2. agreement,

3. fkequently occurring errors due to native language interference, and

- Spelling edors are not treated winithin the system; instead a spelling

qrrection algorithm is applied before the sentence is analyzed by the parser.

The remaining three error classes will be discussed in the following sections.

Agreement

For agreement errors (subject-verb agreement, d e c t e d noun phrase

endings, word order among adverbial elements) the system uses predicates on

the constituents. The predicates check whether the forms are correct. If a

predicate evaluates to fdse, an error message is added. An example of a

predicate might be "CASE-FAILED?". If the predicate evaluates to tme, a

meta-nile is instantiated which loosens the constraint on case agreement.

mThile relaxing constrâints wïü lead to a successfid parse, the strategy

does not necessarily lead to the desired pedagogical outcorne. It is a technique

for permitting ill-formed structures rather than for analyzing 3.l-formedness.

For error-contingent feedback, a system needs to diagnose the source of an

14. Overgeneralization errors are the effects of particular leaming strategies on items within the target language, that is, overgeneralization of target language rules. For example, a shident might choose for the verb schlufen the regular past tense s u f i -te *schlufte as opposed to the correct irregular form schlief. According t o Richards [19711 such learning strategies appear to be universally employed when a leamer is exposed to second language data; this is confirmed by the observation that many errors in second language communication occur regardless of the background language of the speaker. Error Analysis [Corder 1967, Seünker 19721 states that equal weight can be given to interlingual and overgeneralizatio n errors.

Designed Intelligence: A Language Teacher Mode1 14

Page 29: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

error rather than merely relaxhg the constraint. Consider example (3a), cited

by Schwind [1990a, 19951:

(3a) *Der Gom zümea.

(b) Die Gotter Wnen.

me gods are angry.

The subject of the sentence is Der Gotter and thus requires

nominative case. Der Gotter, however, is a e c t e d for genitive plural and the

feedbàck would be me determiner is incorrectly inflected for case. However,

the mistalre is ambiguous and thus depends o n the context.15 In example (3a),

it is very unlikely that the student constructed a subject assigning genitive

case which is the least cornmon a d thus hardest of the four cases in German

[Schwind 19951. The desired feedback should be The determiner is incorrectly

inflected for number since Gotter is plural and der is singular. This analysis is

not possible by simply relaxing the c o n s t r k t on case. It is the context in

which the particular error occurs which needs to be considered t o achieve the

desired feedback.

Errors like the one given in example (3a) cause because the

analysis performed by the parser is correct from a computational point of view

but the feedback is pedagogically wsound. The parser and grammar will not

dag der G ~ e r as an error in number because the determiner and the n o u n

make up a correct noun phrase in the genitive plural. However, when the verb

combines with the noun phrase a case error is recorded because the verb

requires nominative case.

15. Articles in Geman are ambiguous. Depending on where the e r o r occurs der, f o r example, could be a mistake in case, gender, o r number.

Designed Intelligence: A Laquage Teacher Model 15

Page 30: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Frequently O c c h g E m r s due to Native Language Interference

For fkequently occurring errors in word order due to native language

interference, Weischedel et al. Cl9781 add each incorrect form as a buggy rule.

An example of a native language interference error concerns the position of

the past participle in Geman, illustrated in example (4a).

- . --

(b) ~r && das Kind -. He has seen the child.

ATNs do not allow for a very general treatment of errors, particularly

errors in word order. Each individual error has to be anticipated and a buggy

rule needs ta be implemented. Overgeneralization errors in word order, for

example, although not implemented in Weischedel's system, would need to be

treated in a similar way as language transfer errors in ATNs, that is, by ad

hoc d e s . 16

Similar failings, due to the need to anticipate errors can be found in

Liou [1991]. That ATN system consists of an expert model, containing correct

grammatical constructions and a bug model which anticipates the ill-formed

structures. Liou's program covers seven types of errors: det-noun phrases,

conjunctionç, verb morphology, subject verb disagreement, capitalization,

although ... but, and no ngttez.. combinations. The error scope is fairly small

and an extension would involve increasing the bug model by painstakingly

encoding a buggy d e for each likely error

16. Larsen-Freeman & Long L1991J list examples for errors due to factual misconception which they label simplification and induced errors In addition there are mistakes as opposed to errors, a distinction made by Corder 119671. Mistakes are random performance slips. In such cases, the student is aware of the lexemes and gramrnar rules of the target language but due to a lack of Sping skills, concentration, and/or fatigue mistakes occur. Al1 of these error types would require ad hoc rules.

Designed Intelligence: A Language Teacher Mode1 16

Page 31: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Another example of the same heuristics is found in ALICE-chan, a

multimedia foreign language leaming environment for ~ a ~ a n e s e ' ~ developed

by Levin & Evans [1995]. The system covers grammatical topics typically

taught in a fitçt-year Japanese course and includes about 100 d e s . The

limitations are again that speaal rules need to be designed to parse errorful

structures. Thus their treatment is n o t very general.

The h a 1 error class Weischedel addresses is ambiguous readings.

Accordhg to Weischedel et al. [1978] the intended meaning in example (5) is

The man is giuing the girl a hat rather than The girl is giving the man a hat.

(5) Dem Miidchen gibt der Mann einen Hut.

neuter, dative rnasc, nom musc, accusathe

The heuristic Weischedel's system uses t o determine the

interpretation intended by the student is ta select the parse that gives the

fewest errors. Ambiguous readings are rarely addressed by ILTSs although

they are a common phenornenon with highly idected languages such as

~erman. l8

Weischedel's method is a computationally effective way to resolve

ambiguous readings. However, from a pedagogical perspective "a more

sophisticated routine w o d d take into account the likelihood of various kinds

of e r ~ o r s " ~ ~ . The likelihood of an error is determined by the leamer level and

the frequency of a grammatical construction. Consider example (6):

17. The system also includes a Spanish module. 18. Levin and Evans [1995] resolve ambiguity through interaction with the user Covington & Weinrich [1991] use the same heuristics as Weischedel et al. [1978]. However, they express their doubts about the usefùlness of the rnethod employed. 19. Covington & Weinrich [1991], p. 153.

Designed Intelligence: A Language Teacher Model

Page 32: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

(6 1 Sie liebt er.

fem-, nom or fern., acc masc, nom

I f s her he loves.

In example (6) the intended meaning could either be It's her he loves

o r Ske loues him. According t o Weischedel's approach, the former

interpretation would win since the sentence contains no errors, while the

latter has an error associated with the direct object. However, a beginner

student of German most likely intended to write she loves hin, not knowing o r e

overlooking the fact that lieben assigns accusative case to its object. The

interpretation It's her he loues is c e r t d y grammatical if the object receives

stress. But in writing, even native speakers would probably underline Sie

indicating the emphasis of the direct object. An even more puzzling example is

given in (7):

(7) * Sie dankt e r.

fern., nom or fem., acc musc, nom

It's her he thanks or She thanks him.

Choosing the parse with the fewest errors, would not be sufncient for

example (7). The verb danken assigns dative case. The two possible readings

are either It's her he thanks o r She thanks him but in both interpretations

there is one mistake: either sie, nominative is used uistead of ihr, dative o r er,

nominative instead of ihm, dative, respectively In such an instance, relying on

the parse with the fewest errors makes the choice of parse arbitraq even

though the student most likely intended Sie dankt ihm. A better approach to

treating ambiguous readings will be provided in Chapter 3.

Augmented Transition Networks are limited in their application to

ILTSs. In procedural programs, the data and its implementation are

interwoven which makes a system less general and makes it more diffcult t o

provide a large coverage of emors, because errors need to be anticipated and

encoded similar t o a pattern-matching rne~hanisrn.~~

Designed Intelligence: A Language Teacher Mode1

?

Page 33: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Furthermore, alI the ATN systems desaibed above are tied to a

specSc native language of the student, in Weischedel's system, for example,

English. For students with different native languages, the systems wodd

require even more buggy ni les since the error anticipation is derived from the

native language of the student. Examples (8) - (10) illustrate the problem:

(8) German: Ich & es fesehert.

(9) French: Je l'd ou. . -

(10) ~ n & s h : 1 have - it. In German, the direct object appears in between the auxil iq and the

past participle, the auxïiiary always being in second and the past participle in

final position in main clauses. A French leamer of German is likely t o commit

errots concerning the position of the auxiliary, while the EngIish leamer of

German will have problems with the position of the past participle. For an

ATN to deal with the error caused by either native language, buggy d e s of

each er ror would have to be implemented.

A further shortcoming of procedural systems is that they are not

modular. Such systems cannot be easily altered and applied to another

Ianguage. The whole system has t o be rewritten, when preferably one could

replace only the language-dependent grammar withi-n a language-

independent shell.

In distinction tg Augrnenkd Transition Networks, declarative

systems contain feature grammar formalisms which are based on unification.

In addition to meta- and buggy rules, ill-formed input can be processed by

20. Johnson's evaluation of generd ATN systems can be well applied to ILTSs: 'Zike any procedrval model, an ATN gives extensive possibilities for optimization and tight control over details of parsing strategy. The pnce to be paid in return is a danger that as the system increases in size, its logic becornes obscure, modification progressively more risky, and debugging more and more problematic." Johnson [1983J, p.72. See also Loritz 11992, 19951 who M e r discusses the implemenbtion o f a s in ETSs.

Designed Intelligence: A Langaage Teacher Model 19

Page 34: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

altering the unification algorithm itself. Section 1.3 will discuss ILTSs

employing declarative formalisms.

1.3 Intelligent Language Tutoring Systems Using Declarative Formafisms

. * *

The bulk of ILTSs are implemented as declarative representations.

The main types of grammar formalisms used in ILTSs are Government

Binding [Chomsky 1981, 19861 and Logic Grammars (Colmsrauer's

Metamorphosis Grammar Cl9781 and De=te Clause Grammar Pereira &

Warren 19801). In addition to the main types, KPSG as part of the GPSG

family [Gazdar et al. 1985, Gazdar 19821 has been implemented in an ILTS by

Hagen [1994].

There are three common methods of analyising ill-formed input used

in declarative-based ILTSs. In the f k s t approach, in addition t o the target

grammar, the student's native language is explicitly encoded to cover errors

due to native language interference. Depending on the error scope of the

system, meta-des may be added for overgeneralization errors. The second

approach does not rely on the native language of the student, instead relying

on meta-des t o relax the constraints of the target language. In the third

approach, the iinification algorithm is altered in such a way that the system,

despite clashing features, performs a parse and keeps track of the conf l ict ing

features.

An example of the fkst approach is Schuster's [1986] system for

Spanish Ieamers of English. The system focuses o n the acquisition of English

verb-particle and verb-prepositional phrase constructions. The program

Designed Intelligence: A Language Teacher Mode1 - .-

Page 35: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

contains two grammars, Spanish and English. An obvious shortcoming of the

system is that it can only deal with native language interference errors.

Overgeneralization errors, another significant source of students' errors, are

ignored and cannot be recognized by the system.

Catt & Hirst [1990] overcome the shortcoming in Schuster's system

by implementing meta-des for overgeneralization errors. Their system

&ripsi consists of t h e e grammars: English as the target language, French

and Chinese as the native languages of the leamers. Whenever the input

cannot be recognized by English rules alone, the system applies the rules of

French or ~hineçe.~' Although the system represents an extension of

Schuster's system, nonetheless, an important error class passes unrecognized:

errors in word order can only be detected if due t o native language transfer. A

further, rather apparent, limitation of the authors' approach lies in its lack of

generality. First, the same error in the target language has to be encoded in

each source grammar, French and Chinese. Second, identical grammatical

structures are encoded in all three grammars leading to redundancies.

Wang & Garigliano [1992,1995] designed a Chinese Tutor for English

native speakers. In their design, they attempted to avoid the kind of

redundancies found in Catt and HirçPs system. The Chinese Tutor models

only the fragments of the grammatical rules in the native language which are

different from the corre&onding d e s in the target language. Additionally,

the authors address only those transfer errors j u s s e d by empirical data they

collected. They found that 78% of the errors made by English leamers of

Chinese are due to transfer. The remaining 22% are ignored in their system.

21. It is not quite clear from the description in Catt & Hirst 119901 whether a sentence would actually run through three grammars or whether the students possib1y identie their native language when uçing the program and required switches are set.

Designed Intelligence: A Language Teacher Mode1 21

Page 36: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

The percentage of transfer errors is natwally cIosely dependent on the two

languages ~ o r n ~ a r e d . ~ ~

The deficiencies of the systems employing native grammars are

either poor coverage of errors, as in the case of neglecting enors due other

than to native language interference, and lack of generality The systems also

overtly rely on a particular native language of the user and are thus JEmited

in their scope. This-=limitation becomes more problematic due to the increase

in multi-culturalism and on-line Distance Education over the Internet. Any

overt presumption of the native language of the student is severely

restrictive.

A variation of error handling is found in other dedarative systems.

For example, Kurtz, Chen, and Huang Cl9903 developed a DCG system XTRA-

TE for Chinese learners of English. The authors only use correct grammar

rules with meta-des that relax constraints at multiple levels. Level 1

contains all grammatical constrâints, Level 2 relaxes syntax, Level 3

semantics, and Level4 both syntax and semantics. In addition, for translation

exercises the system performs a literal word-by-word translation. The

authors' system is more general in the sense that they do not anticipate

errors on the basis of the native language of the student. However, the system

cannot handle errors in word order. The only constraints which can be relaxed

are agreement emors, both syntactic and semantic.

A similar pitfall can be found in a system for leaming English

developed by Covington & Weinrich [1991]. The program is written in Gulp,

an extended form of Prolog which includes a notation for feature structures

22. Weinreich 119531 asserted that the greater the m e r e n c e between the two languages, i.e. the more numerous the exclusive forms and patterns in each, the greater is the learning problem and the potential area of interference.

Designed Intelligence: A Language Teacher Mode1 22

Page 37: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

[Covlngton 19941. The system focuses o n agreement, insertion, and omission

errors. Insertion and omission errors are due to a violation of the

subcategorization List of the verb. For example, a student might use an

inherently transitive verb without an object. The authors relax constraints by

implementing a lenient d e for each rule which parses well-formed input.

However, errors in word order pass unrecognized.

- Meta-rulespby dennition, cannot address errors in word order. They

c m on ly relax constraint violations of grammatical features. Ali systems

which handle word order errors anticipate these by implementing buggy

rules. However, in ail cases the result is very user-speciiîc since the systems

rely on a p a r t i d a r native language of the student. The ones which

implement more than one native language lack generality The same error in

the target language is presented in each source language.

The third approach in error detection within declarative systems

makes use of the feature grammar formalism i t s e ~ ~ ~ In declarative

implementations, the grammar makes heavy use of the operation on sets of

featues called unification. Unification-based grammars place an important

restriction on unification, namely that t w o categories A and B fail t o unifv if

they contain mutually inconsistent information [Gazdar & P d u m 1985,

Knight 19891. However, this inconsistent information constitutes exactly the

errors made by second language Ieamers. For example, if the two categories A

and B do not agree in gender a parse will fail. A system designed to

accommodate the errors of second language leamers' requires either a

moditication o f the unification algorithm itself or feature specScations

capable of overcoming the gender restriction.

23. Covïngton & Weinrich 119911 use features t o describe the particular error but their error detection mechanism is based on buggy d e s rather than on the feature grammar formalism-

Designed Intelligence: A 1I;anguage Teacher Mode1

Page 38: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Hagen [1994] developed an ILTS, GrammarMaste< for French as a

second laquage. The program is driven by an object-oriented, &cation-

baçed parser mitten in Hypeflalk. The systern addresses three particular

grammatical constructions: conjunctions, reflexive binding, and dislocated,

missing, and s u p e ~ u o u s constituents.

The goal of the system is to show an implementation of some "thorny

problems of complqx grammatical c~nstruct ions-~~~ Hagen's systern borrows *

analyses fkom a number of grammar formalisms: Head-Driven Phrase

Structure Grammar (HPSG), Generalized Phrase Structure Grammar

(GPSG), and Government Binding (GB). Missing conçtituents are handled

with the SUBCAT feature of HPSG, reflexive binding makes use of the foot

feature principle (GPSG), and superfluous constituents are controlled by

thematic assigrunent (GB).

Hagen relaxes the unification algorithm to block parsing faïiure such

that if the parser udetects contradictory features like [geiMJ and [gefl, it

preserves the features o n the major part of speech (in the case of noun

phrases, the noun) and inserts the contradictory feature o n a minor part of

speech like an article into an error stack dong with a corrective message."25

Hagen's implementation of the prinaples fhm GPSG, HYSG, and GB

d o w for a general treatment of the errors considered. T h e errors do not need

t o be anticipated. If any of the grammar principles does not hold, the

constituent is flagged and an error message is attached. However, Hagen

addresses a very small range of errors; the system also lacks a uniform

syntactical fkamework and seems linguistically rather than pedagogically

r n o t i ~ a t e d . ~ ~ In addition, altering unification implies inserting procedural

24, Hagen C19941, p. 5- 25. Hagen 119941, p. 16.

Page 39: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

techniques in an inherently dedarative algorithm. Thus the andysis becomes

closely tied to its implementation, that is, the grammar only works if

unincation is changed in a certain way

Schwind [1990a, 19951 developed a system for German based o n

Colmeraueis Metamorphosis Grammar [Colmerauer 19781. Her system

covers a broad range of errors: agreement, syntactic and semantic errors. Her

method of analyzing-agreement errors is based solely o n the feature gramrnar

formalism. The class of agreement errors covers errors in gender, number,

person, and case of noun phrases. The system does not anticipate these errors.

~nçiead, each lexeme is specified for every possible grammatical occurrence27

In addition, the unification-based algorithm builds up two sets of feature

structures: "... iiriifv (a,b,r,e) holds whenever r is the result of the unification

and e is the set wnsisting of di the pairs of values for which a and b could not

be uniIied, together with all the symbols contained in the symmetrical

difference between a and b."28 Thus Schwind's defkition of unification m e r s

fkom the usual one 19841 in the sense that the parse does n o t fail

but instead records the elements which do n o t un3y

Schwind's analysis also allows for discriminating among errors that

are phonetidy identical, but differ in their source. These errors occur within

26. Grammatical constructions which are linguistically and computationally complex do not necessady present difEculties for the language learner. For example, the errors second language leamers make with conjunctions are Emited to subjectherb number agreement and case assignment of the subject or verb complement. While conjunctions present a challenge from a computational point of view since i t is of utmost importance for the linguist that the parser displays the correct analysis, for the language leamer i t is the feedback that outweighs the linguistic and computational analysis [Farghali 19891. 27. In German, there are three genders, four cases, and two numbers. While this results in 16 potentially distinct lexical forms, there are only six f o m s for definite articles, which are phonetically distinct. For example, the definite article der can be Cgenitive, plural], [genitive, feminine, singular], [nominative, singular, mascl, [dative, feminine, singularl. Schwind [19951 uses only one lexical entry for der, see also lMcFetridge & Heift 19951. 28. Schwind [19951, p. 305.

Desrigned Intelligence: A Language Teachex Mode1 25

Page 40: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

noun phrases. The problem with noun phrases in German derives f?om the

fact that there are three features (gender, nurnber, and case) any of which can

be wrong.29 Consider examples ( l l a ) - (13a), as cited by Schwind [1990a,

19951:

(lia) *Der Kind spielt.

(b) Pas Kind spielt.

nominatiue

The ,&ild is p Zay ing.

(=a) *Er gibt der Milch,

Cb) Er gibt dem Milch.

dative

Ne is giuirrg the child milk.

(13a) * Sie kennt der E n d .

Cb) Sie kemt-.

accusative

She knows the ~ h i l d ? ~

In examples (lla) - (13a), the error occurs with der Kind. According to

Schwind, the desired feedback in example (lla) and (12a) is ZJie de teminer is

incorrectly inflected for gen&r. while in example (13a) it is The deteminer is

incorreêtly inftected for case. In the three examples, although the surface error

der is identical, the sources of the errors are distinct. This is due to the

distinct case requirements of the three examples and the ambiguity of the

determiner del: In example (lia), the source of the error is gender since der is

a possible article in the nominutive (der, die, das). However, &ce der is not a

possible accmative article (den, die, das) as required in example (13a), the

source of the error lies in case. Schwind Cl9953 applies case filtering to achieve

the desired feedback. Her system fmt checks whether the article is a possible

determiner for the required case. If it is, the system responds with a gender

29. These three features do not cover the declension of adjectives. Adjectives also i d e c t according to whether or not they are preceded by a definite o r indefinite article. 30. Schwind C19951, p- 312.

Designed Intelligence: A Language Teacher iMode1 26

Page 41: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

error; if not, the source of the error is case. The shortcoming of her analysis is

that the system assumes that students in general are more likely to know

grammatical case than gender. This is apparent in example (12a) which

Schwind Cl9951 analyzes as an error ingender as opposed to in case. The error

is ambiguous due t o the grammatical ambiguity of der. If the student

assumed that Kind is feminine, the error is due to wronggender. However, if

the student knows the correct gender but did not assign the correct case, the

error-source is wrdiig case. Ambiguous errors will be further discussed in

Chapter 2.

Schwind uses the same analysis for semantic errors. The system

recognizes the violation of semantic restrictions on verbs and their

compIements. For this, she provides a semantic network where semantic

constraints are expressed as feat~res.~ ' For example, while cornputer is

[inanimate], woman is marked as [animate, human]. The analysis is again

very general, that is, the errors do not have to be anticipated.

One sigdicant deficiency of Schwind's system is its handling of high-

level syntax errors. In syntax, Schwind makes a distinction between low- and

high-level. Low-level syntax errors are errors of insertion or omission while

high-level errors refer to errors in word order. Schwind uses one buggy rule

for each constituent of insertion and omission. High-level syntax errors,

however, must be anticipated and according to Schwind Cl9951 their

treatment is not general.

Feature grammars are more promiçing than ATNs for Intelligent

Tutoring Systems. The data and implementation are kept separate which

allows for easier expansion. Feature grammars also accommodate more

31- see also Weischedel et al. [1983]. The authors present a very similar analysis of semantic networks. However, i t is not as complete as the one by Schwind C1990b1.

Designed Intelligence: A Tnnguage Teacher Mode1 27

Page 42: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

methods of error detection. In addition to meta- and buggy-rules, they allow

for alteration of the unification algorithm itself Altering the unification

algorithm has its pitfalls as rnentioned; however, a more general treatment of

errors than with meta- and buggy rules alone can be achieved.

Feature relaxation presents an alternative t o existing methods and is

the method adopted in t b i s dissertation. Unlike existing ILTSs, the analysis

descrïbed does not 'keek or anticipate errors, but instead emulates the way a

language instructor evaluates a student's response. The computational

analysis reflects the pedagogicd bias of the system.

1.4 Evaluation of Intelligent Language Tutoring Systems

The foregoing discussion illus trates t w o developmental stages in

ILTSs that primarily focus on form rather than o n content. The early ILTSs,

as the one by Weischedel e t al. [1978], concentrated solely on tackling the

computational problems of parsing iIl-formed input as opposed t o embedding

pedagogic considerations into such systems. For example, while Weischedel's

system considers ambiguous readings, they are addressed fkom a

computational rather than a pedagogic point of view. This is evident in the

algorithm used to handle such errors. Selecting the parse wi th the fewest

errors is a computationally effective method, but it does n o t address the likely

cause of error. A pedagogic system would consider the performance level of the

learner andior the language leamhg task in order t o address ambiguous

readings.

Desigaed Intelligence: A Language Teacher Mode1 28

Page 43: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

The second phase of ILTSs, while still extending the parsing

capabilities of iu-formed input, also attempts t o emphasize pedagogic

considerations. For example, Schwind [19951 and Holland [1994] seek a wider

error coverage in a computationally more generd way, and also address

language learning pedagogy in c o n s i d e ~ g ambiguous errors. However, their

pedagogic focus is Zimited. While the alg0ntb.m addresses ambiguous errors to

some extent, contingent errors cannot be handled adequately - - #

The discussion of existing lLTSs shows that neither phase has been

completed. From a computational point of view, some errors such as errors in

word order, for example, s t i l l need to be anticipated and thus are not

addressed in a general way. The algorithms are based on a particular native

language of the student and the systems are thus limited in error coverage

and user scope. From a pedagogical perspective, ambiguous and contingent

errors are important to language instruction.

The analysis presented in this dissertation addresses both

developmental phases in ILTSs. From a computational point of view, a wider

class of errors is addressed in a general wax thus n o t s a d c i n g error

coverage and user scope for errors in word order. From a pedagogical point of

view, the andysis considers the learner by emulating the ways a language

instnictor evaluates and responds to a student's performance in language

learning tasks. As a resdt, ambiguous and contingent errors, as well as

multiple sentence readirigs, are treated with a strong pedagogic focus. In

addition, instructional feedback is suited to learners' expertise.

The following section will outline the Gennan Tutor, which austrates

the analysis desrribed in this dissertation by way of example.

- -

Designed Intelligence: A Language Teacher Mode1

Page 44: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

1.5 The G e r m a Tutor

The NLP component of the G e m a n Tutor parses students' answers to

introductory German exertises and returns structures which specdy student

model updates and possible feedback of diEerent levels of granularity - - -

coaelated with çhidents' expertise. The concept of granularity has been

previously applied to an Intelligent Tutoring System for LISP programming

[see Greer & McCalla 19891. For ILTSs, granulari@ is particularly important

in &amhg responses to learners' errors. Inexperienced students require

detailed instruction while experienced students benefit best fkom higher level

reminders and explmations LaReau & Vockell19891.

The design of an ILTS implemented in the G e m n Tùtor consists of

five components: the Domain Knowledge, the Licensing Module, the Analysis

Module, the Student Model, and the Filtering Module. The five components,

given in Figure 1.2 correspond t o the steps a language instiuctor applies in

evaluating and responding to students' errors.

The Domain Knowledge represents the knowledge of the language. It

consists of a parser with a grammar which parses sentences and phrases t o

produce sets of phrase descriptors NcFetridge & He* 19951. A phrase

descriptor is a model of a particular grammatical phenornenon such as case

assignment or number agreement. A phrase descriptor records whether o r not

the grammatical phenomenon is present in the input and correctly or

incorrectly formed.

Due to structural ambiguity found in language, a parser produces

more than one parse for many sentences. The Licensing Module selects a

Designed Intelligence: A Language Teacher Model

Page 45: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Student Input

knowledge of the language encoded in the grammar

. - J + sends phrase descriptors of dl parses

Licensin~ Module

seIects the desired parse

i sends al1 phrase descriptors of the selected parse

retums student model updates and produces instructional

feedback of increasing abstraction

4 sends a student model update and instmctional feedback

l Student Model I l keeps learner model updates

and decides on the student level I 1 sends instructional feedback suited to student expertise

fil ter in^ Module decides on the order of instructional feedback

Error-contingent Feedback Suited to Learner Expertise

Figure 1.2: The German Tutor

Designed Intelligence: A Language Teacher Model

Page 46: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

parse by taking into account factors a language instructor considers in

evaluating a student's response. In determining the most likely sentence

reading, a language instructor considers the level of instruction, the

eequency of the grammatical construction, and thus the likelihood of the

error.

The Analysis Module incorporates a language instructor's knowledge

in piqointing the precise source of an error The Analysiç Module takes a

phrase descriptor as input and generates sets of possible responses to the

leamer's input that the instruction system can use when interacting with the

student. The level of the learner, either expert, intermediate, o r novice

according t o the m e n t state of the Student Model, determines the particular

feedback displayed. The Student Model records mastery of grammatical

stnictures as well as structures w i t h which the learner has problems.

The Filtering Module determines the order of the instructional

feedback displayed t o the learner. The system displays one message a t a time

so as n o t to overwhelm the student with multiple error messages. In addition,

the Filtering Module takes into account contingent errors, a class of multiple

errors t o be discussed in Chapter 4.

1.5.1 The Domain Knowledge

The parser analyzes students' input according t o the knowledge of the

language encoded in the grammar. The parser and the gramrnar thus provide

the linguistic analysis of students' input. The grammar fo r the German n t o r

is written in ALE (The Attributed Logic Engine), an extension of the

cornputer language Prolog. ALE is an integrated phrase structure parsing

and definite clause p r o g r d g system in which grammatical information is

expressed as typed feature structures. Typed feature structures combine type

Designed Intelligence: A Language Teacher Mode1 32

Page 47: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

inheritance and appropriateness specifications for features and their values

[Carpenter & Penn 19941.

The grammar formalism used is derived &om Head-driven Phrase

Structure Grammar Vollard & Sag 1987, 1994, Nerbonne et al. 19941. This

theory is one of a family which share several properties. Linguistic

information is presented as feature/value matrices. Theories in this family

are -to yarying dk-&ees lexicalist, that iç, a considerable amount of

grammatical information is located in the lexicon rather than in the grarnmar

rules. For example, Figure 1.3 illustrates a minimal lexical en- forgeht. The

subcategorization list of the verb, notated with the feature subi, specifies that

geht takes a subject which is minimdy specifîed as a singular noun. Rules of

gra-ar specify how words and phrases are combined into larger units

according to the suhcategorization List. In addition, there are principles which

govern how information such as the head features is inherited.

phon < geht z

vat

head tt

subj cat [head n]

content kndex [num

Reln geht [Geler 1

Figure 1.3: Partial Lexical Entry forgeht

Designed Intelligence: A Isnguage Teacher Mode1 33

Page 48: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

The philosophy of the analysis described differs fkom other systems

designed to parse ill-formed input in ILTSs in that it does not seek or

anticipate errors, but instead records whether or not grammatical constraints

are met, the goal being to analyze students' language input rather than the

more modest aim of recognizing ïü-formed constmctions. The analysis rnodels

the steps taken by a language instructor in the correction process of students'

input. - - f-

The grammar itself is written as a set of constraints. Any constraint

such as the one in Figure 1.3 that the subject of the verb geht must be a

singular noun may block parsing ifnot successfully met. The terminology used

in Figure 1.3 will be described in more detail in Chapter 2.

In HPSG, the information flow is modeled by structure-sharing and

not by transformation o r movement. Two o r more distinct attributes (or paths)

within a feature structure are said to be structure-shared if their values are

specitied by one and the same feature structure Pollard & Sag 19871.

Structure-sharing in feature-value matrices is indicated by multiple

occurrences o f a coinderring box labeling the single value. For example, in the

lexical entry for the verb geht, given in Figure 1.3, the number feature of the

verb and its subject share one common value, x. Ifthis condition is not met,

that is, either the subject o r the verb is not singular, unification fails. In ILTSs,

however, sentences contghing such inconsistent information need to parse

successfully so that the error can be discovered, analyzed, and reported t o the

student.

The constraint on number agreement c m be relaxed, however, by

changing its structure so that, rather than checking that the noun is singular,

the system records whether or not the subject of geht is in the singular. To

achieve this, the noun is no longer marked as [num d, but instead the path

Designed Intelligence: A Language Teacher Mode1 34

Page 49: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

numl sg terminates with the values ermr or çorrect- For example, for a

singular n o u n phrase, the value of the path num I sg is correct, while it is error

for a plural noun phrase. The two partial lexical entries are given in Figure

1.4 and Figure 1.5, respectively

-

2- synsem local content index I

Figure 1.4: Markhg Number Featmes for Singular Nouns

.-

synsem

C

local content

Figure 1.5: Marking Number Features for Plural Nouns

The verb geht records the value of ~g hom its subject (Figure 1.6). If

t h e value of the path n v I sg is correct, the subject is in the singular. In case

of a plural noun, geht records the value error for number agreement.

The goal of the parser and the grammar is the generation of phrase

descriptors. A phrase descriptor is implemented as a fiame structure that

models a grammatical phenornenon. Each member of the fiame consists of a

name followed by a value. For example, subject-verb agreement in number is

modeled by the Game [number, value] where value represents an as yet

Designed Intelligence: A Language Teacher Mode1 35

Page 50: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

phon < geht >

local

cat * - i

head v

subj

des criptor

synsem 1 cat [head n]

local

Figure 1.6: Recording Number Features forgehf

uninstantiated value for number. If the grammatical phenornenon is present

in the student's input, the value is either correct or error depending o n

whether the grammatical conskaint has been met or not, respectively. If the

grammatical constraint is missing, the feature value is absent. Consider

examples (14a) and (b):

6) Er geht.

He is leaving.

The phrase descriptor for subject-verb agreement in number in

example (14a) is [numberyerrorl, while that for the sentence in (b) is

[number,correct]. For either sentence, (14a) o r (b), the information will be

recorded in the Student Model. A system presented with (14a), however, will

also instnrct the learner on the nature of subject-verb agreement in number.

Designed Intelligence: A Language Teacher Model 36

Page 51: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

In addition to the grammat ica l features defined in HPSG the

grammar uses a type descriptor representing the description of the phrase

that the parser builds up. This type is set-valued and is initially

underspecified in each lexical entry- During parsing, the values of the features

of descri~tor are specified. For example, one of the members of descriptor,

vp num in Figure 1.6, records the number agreement of subject-verb in a

main-clause. Its value is inherited from the feature speded in the verb - -

geht. Ultimately, descnotor records whether the sentence is grammatical and

what errors were made.

1.5.2 The Licensing Module

The Licensing Module determines the most likely reading of a

sentence. Due to language ambiguity. parsers generally produce more than

one syntactic structure. Feature relaxation leads to an even greater number o f

parses. Selecting the fkst parse o r the parse with the fewest errors is

inappropriate for the design of an ILTS described. Neither solution takes into

account the likelihood o f an enor, possibly resdting in misleadhg feedback.

The design of the Licensing Module reflects a language instructor's decision in

selecting the most likeIy reading of a sentence by considering the source

language of the learner. the level of instruction, and the fiequency of a

grammatical construction. This is discussed in detail in Chapter 3.

1.5.3 The Analysis Module

The third component of the system is an Analysis Module which

encodes knowledge of the language instructor to determine the exact source of

an error. The Analysis Module takes phrase descriptors as input and

generates possible responses that the instruction system can use when

Designed Intelligence: A Language Teacher Model 37

Page 52: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

interacting with the leamer. A response is a pair that contains a message the

system will use to inform the leamer if a phrase descriptor indicates there

has been an error and a Student Model update. The Student Model update

contains the name of a grammar constra.int in the Student Model dong with

an instruction to increment or decrement the corresponding cumulative total.

The Analysis Module generates sets of instructional feedback of

increasing abstraction. As an example consider the ungrmatical sentence

in (15a). An inexperienced student should be informed that Madchen is a

neuter noun, that the verb danken is a dative verb and that the determiner

das is incorrect. A student who has mastered case assignment (as indicated

by the Student Model) may be informed only that the case of the object is

incorrect.

(15a) *Der Mann dankt m. (b) Der Mann dankt

TILE man t h a h the girl.

The Analysis Module is implemented in DATR Evans and Gazdar

19903, a language designed for pattern-matching and representing multiple

i n h e r i t a n ~ e . ~ ~ The Sussex version of DATR is implemented in Prolog.

Nodes in DATR are represented by the name of the node followed by

paths and their values. A partial representation of the node which

corresponds to the phrase descriptor that records the position of a finite verb

in a main clause is given in Figure 1.7.

The paths in a node definition represent descriptions of grammatical

constraints monitored by phrase descriptors. The matching algorithm of

32. The andysis makes minimal, although essential, use of the multiple inheritance capabilities of DATR.

Demgned Intelligence: A Language Teacher Mode1 38

Page 53: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

<main,clause position-mainclause finite errer> ==

Cposmainclfin' '3' ' m e ' 'The verb in the main clause is not in the correct position.'

'posmainclfin' '2"true1 'The finite verb in the main clause is not in the correct position.'

'posmainclfin' 'S 'true' 'The finite verb in the main clause is not in the correct

position. I t has to be the second element of the sentence.')

<main-clause positionmaincIatuse finite correct> == ('posmaincEnl ' 1' 'false' "

'posmainclfin' '1' 'fdse' "

'posmainclfin' '1' 'fdse' ")

<main-clause position,mainclause finite absent> = (").

Figure 1.E DATR Code Listing for a Finite Verb in a Main Clause

DATR selects the longest path which matches left to right. Each path in a

node is associated with atoms on the right of '=='. There are three learner

levels considered: expert, intermediate, and novice each of which has four

a t o m ~ . ~ ~ The first atom in each set refers to a gmmmar node, followed by an

increment, a Boolean value, and a message?

For example, the node gosrnainclfin, is responsible for monitoring

finite verb position in a main clause. If there has been an error identified by

the parser, the parse will generate the phrase descriptor [mainainclause

[positi~n~mahclause [fùiite errer]]]. This will match the path anain_clause

position-mainclause h i t e errorx This clause will concatenate the name of a

d o t in the Student Model pusnainclfin and its i n c r e ~ n e n t ~ ~ with a message.

The message provides the student feedback, and the remahhg information is

used to adjust the Student Model.

33. A phrase descriptor tha t contains the value correct indicates that the grammatical constra.int has been met in which case no feedback message is specified. A phrase descriptor that contains the value absent is ignored by the system. Thus the Est to the right of '==' is empw- 34. The Analysis Module will be explained explicitly in Chapter 4. 35. The Boolean values indicate whether it is an increment or decrement: true for increment, false for decrement.

Designed Inteiligezcs: A Language Teacher Mode1 39

Page 54: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

If the phrase descriptor is [mainclause [position-mainclause [n i te

correct]]] and thus there has been no error, no message is associated with the

name of the slot in the student model and its demement. The information,

however, st i l l contributes a Student Model update in order to record the

success.

Finallx if the grammatical phenornenon was not present in the

student input, the phrase descriptor [main-clause [position-maindause [finite

absent]]] is generated. However, the phrase descriptor is uninstantiated and

thus no information is associated.

The &al result of the Analysis Module is a student model update and

a list of possible learner responses fiom a coarse t o fine grain size. The

information is further processed by the Student Model.

1.5.4 The Student Model

The Student Model keeps track of the learner history and provides

l e m e r model updates. There are three learner levels considered in the

system: novice, intermediate leamer, and expert. The Student Model passes

instructional feedback suited t o learner expertise to the Filtering Module.

The Student Model consists of 79 grammar constraints which are

equally distributed among main, subordinate and coordinate clauses.35 The

distinction between the three clause types is necessary for two reasons. First,

if not kept distinct the system would overwrite the descriptors instantiated by

one clause type once it parses the next. Second, in some instances, there are

35. Each clause contains 25 m m a r constraints per clause. In addition, there are two grammar constraints monitoring noun phrases practiced in isolation, one for verb-initial position, and one for the entire sentence. A complete Est of the grammar constraints is provided in the Appendix.

Designed Intelligence: A Langaage Teacher Model 40

Page 55: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

difXerent grammatical rules which apply to main and coordinate as opposed to

subordinate clauses. For example, in a subordinate clause in German the

finite verb appears in final position while in a main clause it is always the

second constituent. To capture this distinction and to provide error-contingent

feedback suited to learner expertise the system needs to i d e n m in what kind

of clause the constraints were met o r violated.

- Each of the grammar constraintç represents a grammatical

phenornenon; the phrase descriptors report which constraints have been met

o r not. Each student starts out at the intermediate level. Once a student uses

the system, the Student Model adjusts the counter of each grammar node

accordingly. Setting the initial level to intermediate is arbitrary but

reasonable, and the model will quickly adapt for various profiles.

On the basis of the dynamic leamer model, the system decides on the

granularity of the instructional feedback. The central idea is that an expert

requires less detailed feedback than a beginner leamer Fischer & Mandl

19881. For each gramma. node, the Student Model checks which level the

student has achieved and sends the appropriate instructional feedback to the

F'iltering Module.

1.5.5 The Filtering Module

The 6nal component of the design is the Filtering Module which

decides on the order of the inçtmctional feedback displayed t o the learner. In a

language teaching situation a student might make more than one error mithin

an exercise. In a typical student-teacher interaction, however, the language

instructor does n o t overwhelm the student with multiple error reports.

Instead, a language instructor reports one error at a time by considering the

salience of an error. This pedagogy has been transferred t o the analyçis of this

Designed Intelligence: A Language Teacher Model 41

Page 56: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

dissertation. The grammar constraints produced by the phrase descriptors are

hierarchically organized. An E r o r Priority Queue determines the order in

which the instructional feedback is displayed by considering the importance of

an error in a given exertise and the dependency between syntactically higher

and lower constituents. Once the student makes the required correction, the

whole evaluation process repeats.

- C~ntingent~errors , a class of multiple errors which presents a

particular problem for existing systems, are taken i n t o consideration. For

instance, example (16a) illustrates that denken subcategorizes for the

preposition an, while von is a dative preposition requiring the dative pronoun

dir. Ignoring the dependency of errors, the feedback to the student would be

Titis is the wrong preposition and This is the wrong case of the pronoun. The

feedback is correct with regards to von, however, the pronoun dich is not

incorrect if the student correctly changes the preposition von to an. Depending

o n the order of the feedback, the student might end up changing von to an and

dich to dir and wind up utterly confused because the case of the pronoun had

been %agged as an error in the original sentence.

U6a) * Ich denke - von dich.

dat prep- acc pronoun

6) Ich denke - an - dich.

acc prep. acc pronoun

I am thinking of you.

The error in the pronoun is correctly ffagged by the system fkom a

purely logical point of view. However, f?om a pedagogical perspective reporting

the error in the pronoun is redundant and even misleading- In such instances,

only the error in the preposition is reported while the error in the pronoun is

recorded in the Student Model. Once recorded in the Student Model the

information can be used for assessing and remediating the student.

Designed Intelligence: A Language Teacher Mode1 42

Page 57: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

ILTSs aim at modelling the human tutor. Thek goal is to provide

error-contingent feedback and to allow for an individualization of the leaming

process. From a computational point of view, the challenging task lies in

pro cessigg iIl-formefi input.

The existing systems discussed all depend o n buggy and meta-rules,

and/or an alteration of the dcat ion-based algorithm itself. Even in the most

general systems, errors in word order need to be anticipated and thus cannot

be treated in a very general wap. The search for particular errors is also to a

large extent based on the native language of the student. Any system which

anticipates errors requires that the same error in the target language be

presented in each dinerent source language. The fewer source languages

considered the more user-limited the system. In addition, in a number of

systems the data and its implementation are interwoven complicating their

design, maintenance, and ultimately restricting their error coverage.

The strengths of the analysis of this dissertation are manifold: first,

the design contains a pedagogic component and thus it emdates a language

instructor by evaluating and responding to students' language input. The

result of the entire process is error-contingent feedback suited to learner

expertise. Second, the grammar is sufncient-y general that it treats

grammatical and ungrammatical input identically for the phenornena it is

designed to handle. This generality has the advantage of reducing the number

of rules required by the grammar. Third, the decoupling of the parsing system

from the analysis of whether or not the input is grammatical has the practical

advantage that development of each can proceed independen* Fourth, the

Designed Intelligence: A Language Teacher Mode1 43

Page 58: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

analysis described can successfùlly handle ambiguous and contingent errors,

a capability beyond that found in existing systems.

The modular design of the entire system also allows for easier

adaptation to other languages. Although the grammar itself is language-

dependent, the system does no t anticipate specinc errors and thus is not built

entirely fiom the perspective of a particular native language. The Licensing

Module, the Analysis Module, the Student Model, and the Filtering Module /

are langjuage-dependent only to a trivial degree. While the phrase descriptors

and thus the instructional feedback will Vary hom one language to another

the overd architecture of these modules can be applied to languages other

than Gennan.

Designed InteUigence: A Language Teacher Mode1 44

Page 59: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

C H A P T E R 2 Analyzing Student Input

2.1 Introduction

The f i s t step in evaluating students' input is sentence analysis. In

ILTSs, this is handled by the parser and the grammar. The parser analyzes a

sentence according to linguistic information encoded in the grarnmar.

In Head-driven Phrase Structure Grammar (HPSG), linguistic

information is formally r-epresented as feature structures. Feature structures

spe- 'values for various attributes as partial descriptions of a hguistic

sign. For example, the feature structure aven in Figure 2.1 provides partial

information about the verb geht.

According t o Pollard & Sag [1994], each linguistic object is assumed t o

have at least the attributes Phonology (phon) and Syntax-Semantics

(synsem), and in the case of phrases, Daughters (dtrs). The phon value

Designed Intelligence: A ILanguage Teacher Mode1 45

Page 60: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

1 phon < geht >

1 L

Figure 2.1: Partial Feature

[cat [head n]

content 1 index l

Structure forgeht

'1

specines the phonetic f o m of the sign; the svnsem value provides syntactic

and semantic information; and the dtrs value describes the constituents of

the phrase.

HPSG adopts a lexicalist approach in which syntactic information is

described within each lexical en- Subcategonzation, a specification by the

head for its argunent(s), is represented by the features (specïfier), subi - .

(subject), and comr>s (complements). For example, the feature stmcture given

in Figure 2.1 illustrates that the verb geht subcategorizes for a subject. The

subject is minirilally specified as a noun and its person and number features

are stnicture-shared with the agreement features of the verb. Structure-

sharing is indicated by multiple occurrences of a coindexing box labelhg the

single value. The comns Est, which refers to the number and kind of

complements the verb takes, is emp@ since geht is an intransitive verb.

Designed Intelligence: A Langnage Teacher Mode1 46

Page 61: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Because syntactic information is expressed within each lexical entry, HPSG

requires only a few generalized syntactic d e s to spec* how words and

phrases combine into larger units.

In unification-based grammars, grammatical phenornena are

modeled by feature structure matrices. The analysis in this dissertation has

adapted the syntax/semantics component of HPSG to the error analysis

student input twugh the concept of phrase descriptors. A phrase

descriptor is a mode1 of a grammatical phenomenon that records whether o r

not the grammatical phenomenon is present in the input and, if so, whether it

is correctly or incorrectly formed. In either case, detailed information is

collected, which can then be passed o n ta other program modules for further

analysis, and dtimately used to generate appropriate feedback.

A phrase descriptor is implemented as a fiame structure. Each

member of the frame consists of a feature followed by a value. For example,

subject-verb agreement in person is represented by the & m e [per, value]

where value represents an as yet uninstantiated vdue for person. If the

grammatical phenomenon is present in the studentys input, the value is either

correct or error depending o n whether the grammatical constraint has been

met or not, respectively. If the grammatical constraint is misskg, the feature

value is absent. To examine this concept, consider examples (la) and (b):

(la) *Er schlafst.

(b) Du schlafst.

You are sleeping-

The phrase descriptor for subject-verb agreement in person in (la)

would be [per, errorl, while that for the phrase in (b) wodd be [per, correctl. In

either instance, the error o r success will be recorded in the Student Model.

Presented with (la), however, the system will additionally inform the learner

of the error.

Designed Intelligence: A Language Teacher Mode1 47

Page 62: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

A grammar conçists of a lexïcon and syntactic rules. For an ILTS,

there are thus two places for recording whether grammatical constraints have

been met: within the leicon or within the grammar rules. The difference lies

in the suitabdi@ of each location in application to a particular type of error.

For instance, under the design I am describing, errors are grouped into two

main classes: errors in feature matching and errors in linear precedence.'

Errors in feature matching are treated within the lexical entries; errors in

h e & precedence a& addressed in the grRmmar rules.

Most of the errors made by introductory students of ~ e r m a n ~ can be

analyzed by feature rnatching. Examples of such errors are gender, number

and/or case mismatch between two or more constituents. These are found in

noun phrase concord, o r in the agreement between a preposition and its NP

argument, for example. In addition t o violation of agreement of syntactic

featues, a violation of agreement of semantic features also f d s i n t o this

class.

Errors in word order are identified withjn the grammar rules. One

conçtra.int o n word order in German is verb position. For instance, there are

three possible positions for finite verbs: initial, second, and last. Examples (2)

and (3) illustrate that the f i t e verb occurs in initial position in both

questions and imperatives. It appears in h d position in embedded clauses

(example (4)) and in all-other instances it is the second constituent of the

sentence (examples (5),(6),(7),(8)).

(2) Sitzt Klaus im Garten?

Is Klaus sitting in the garden?

1. This classification is language-independent. 2. Studies [Nemark 1976, Rogers 1984, J u o d y n a s 19941 have identified agreement phenornena as a major source of errors for students of German.

Designed Intelligence: A Language Teacher Model 48

Page 63: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

(3) G& in den Garten!

Go into the garden!

(4) Ich weiss, dass Klaus im Garten m. I know that &?am is sitting in the gariien-

(5) Wem g& du das Buch?

To whom are you giuing the book?

(6) Hlaus && im Garten. . -

Kluug is sitting in the garden.

(7) Heute sitzt Klaus im Garten.

irbday maus is sitting in the garden-

(8) Dass maus irn Garten sitzt, weiss ich.

That =arts is sitting in the garden, 1 know.

For nonfinite verbs, there are t w o possible positions: the nodinite

verb occurs in &al position in main clauses (example (9)) or second t o last in

embedded clauses (example (IO)), the last element being the h i t e verb

(example (10)).

(9) Khus hat im Garten -. dSaus mas sitting in the garden-

(10) Ich weiss, dass Klaus im Garfxn m- hat.

1 know t h t maus was sitting in the garden.

The following discussion will focus o n the design underlyhg the

generation of phrase descriptors for errors in feature matching and errors in

linear precedence. Unlike other ILTSs, the analysis presented does n o t

anticipate errors, a problem discussed in Chapter 1. Instead the parser -. analyzes a sentence by comparing input against the knowledge of the

language spedied in the grammar. A phrase descriptor records a grammatical

phenornenon present in the input by describing its contextual occurrence.

Designed Intelligence: A Language Teacher Mode1

Page 64: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

2.2 Errors in Feature Matching

Errors in feature matching are handled within the lexicon by

adjus ting how grammatical phenornena are represented in lexical entries. The

following section will explain the approach for recording student performance

on eriors in feature-matchhg. The algorithm will be illustrated with subject-

verb agreement in person.

2.2.1 General Technique for Errors in Feature Matching

To illustrate the approach of errors in feature matching, the

discussion in this section will focus on the two sentences given in example

(lla) and (b):

(lla) *Er g&&.

(b) D u m .

You are leauing.

The sentence in (lla) contains a mistake in person agreement: er is a

3rd person, singular, nominative pronoun while gehst is a 2nd person,

singular verb.

A grammar is written as a set of constraints. In example (lla), the

constraints between the subject er and the verb gehst require agreement in

number, person, and case. Any of the three constraints could block parsing if

not successfully met. Figure 2.z3 illustrates that the subject of the verb gehst

must be a 2nd person, singular, nominative, noun. Er, however, given in

3. The lexical enties illustrated in this chapter contain only partial feature specifications for breviw

Designed Intelligence: A Language Teacher Model 50

Page 65: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

cat .

head v

local

- phon < geh

Reln gehst 'ont"t[Geher 1

synsem

-.

f

1

-.

cat [head ~ m s e nom.]

content [index

d

2

d

-

1

Fignre 22: Lexical Entry for gehsf

Figure 2.3, is intlected for a person, singular, nominative. Thus *Er gehst

local cat [head ,[case nom]] 7 p&{

Figure 2.3: Lexical Entry for er

A constraint can be relaxed, however, by changing its structure so

that, rather than enforcing agreement between two constituents. it instead

Designed Intelligence: A Language Teacher Model 51

Page 66: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

records whether or not agreement is present. To achieve this, the subject er,

given in Figure 2.4, is no longer marked as [per 3rd1, [num &, [case nom].

Instead the feature gt% for example, specifies all possible person features,

that is, 1st. 2nd. and 3rd. For er, the value for and 2nd is error, while for

3rd it is correct. Conceptually, the feature structure States that it is incorrect

to use er as a 1st and 2nd person pronoun, but correct for 3rd person.

cat ihead n]

Figure 2.4: Marking Person Features for er

The verb gehst, given in Figure 2.5, no longer subcategorizes for a

subject marked [per 2ndl. Instead, the verb gehst will inherit the value of 2nd

fiom its subject during parsing. For the subject er, the value for 2nd is error

indicating that the constraint on person agreement has been violated, but,

importantly, allowing the parse to succeed.

For the correct sentence D u gehst, given in (b), the value for 2nd

would be correct because du is a second person, nominative pronoun, as

iIlustrated in Figure 2.6. During parsing, gehst will inherit the value correct

indicating that the constraint on person agreement has been met.

In addition t o the HPSG features, the grammar uses a feature

descrintor representing the description of the phrase that the parser builds ~~~~~ -

Designed Intelligence: A Language Teacher Model

Page 67: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

phon < gehst >

head v

subj synsem local

content index per 2nd 2 [ [ q R e h gehst

lGeher

lescriptor

Figure 2.5: Marking Person Features for gehst

synsem local

cat [head n]

content index I

Figure 2.6: Markïng Person Features for du

Designed Intelligence: A h g u a g e Teacher Mode1 53

Page 68: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

up. During parsing, the values of the features of descri~tor becorne ~ ~ e c i n e d . ~

For example, the phrase descriptor v p q e r records the constraint on person

agreement. For the sentence *Er gehst, v p g e r will inherit its value f?om the

feature 2nd, given in the lexical entry ofgehst in Figure 2.5. The nnal result is

the phrase descriptor [mainclause [vpger [Znd errorl]], indicating that the

required grammatical constraint on person agreement has not been met.

- The responpibility of the head of a phrase is t o collect the phrase

descriptors of its ~orn~lernents .~ In the implementation in this dissertation,

the phrase descriptors are percolated up the syntactic tree via the Deschptor

Prtnciple. This principle states that the descriptor features of the mother node

are the descriptor features of the head daughter.

The advantages of the method described for analyzing constraints in

featue matching are:

First, the technique is very general in that it can be applied to any

agreement phenomenon. The phrase descriptors record whether a

grammatical phenomenon is present in the input and, if so, whether it is

correctly o r incorrectly formed.

Second, errors are not treated differently from well-formed input.

Thus, errors in feature matching need not be anticipated. Neither the parsing

nor the grammar formalism needs to be altered and no additional d e s are

required.

4. takes three features: main cocoordinate c h , and sub clause. These record the constraints within the three clause types. Each of these in turn, contains a number of features. 5. This knowledge must be specified in the lexical entries as opposed to the grammar d e s . For errors in word order, for exampie, a head inherits phrase descriptors from its complement which in many caseç are not instantiated until afker the parse. To obtain the instantiated values, the lexical entries define which phrase descriptors a head d l inherit fkom its complement.

Designed Intelligence: A Language Teacher Mode1

Page 69: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Third, the method used tc idenfi@ errors in feature matching can also

handle ambiguous errors, and therefore represents an improvement over the

existing techniques. The method closely modeIs the reasoning a language

instructor employs in disarnbiguating enors. Ambiguous errors will be

discussed in the following section.

2.2.2 Ambiguous Errors

Ambiguous errors are distinct f?om ambiguous readings, discussed in

Chapter 1. Ambiguous readings are due to stnictural ambiguity in the

language and manifest themselves as multiple parses. For these, the task is to

select the appropriate parse among many parses. Ambiguous errors, however,

do not necessarily result in multiple parses. They are ambiguous with regard

to the source of the error. The task here is to iden* the most likely error so

that appropriate feedback c m be generated.

Ambiguous errors are those whose source must be identified f?om a

range of possibilities. An intelligent system must emulate the techniques a

language instructor employs in disambiguating errors. The rationale a

language teacher uses is based on the Lkelihood of an er ror as determined by

the student level, previous performance history, and the frequency and

dinicuky of a particular grammatical construction [see Rogers 19841.

The following discussion will provide examples illustrating

ambiguous errors. The example given in section 2.2.2.1 will show an instance

in German where a parser might flag o n l y one of t w o possible sources of an

error. In such cases, however, a language instmctor will typically address both

reasons for the error so as to not misguide the student. Section 2.2.2.2 will

give an example where a parser might flag a n unlikely error resulting in

feedback which is pedagogically unwarranted. In section 2.2.3, 1 will show

Designed Intelligence: A tanguage Teacher Mode1 55

Page 70: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

how the method for errors in feature rnatching can emulate a language

instructor in analJrzing both kinds of ambiguous errors. The method employed

provides a more general analysis than existing systems.

2.2.2.1 Two Likely Sources of an Error

Ambiguous errors have more than one possible reason for the error.

There are instances, however, where the feedback will be misleadhg if the j-

syçtem reports only one of the potential errors. For example, in sentence (IZa),

given by Holland [1994], the student made a mistake with the determiner die

of the prepositional phrase. The source of the error is indeed ambiguous: it is

either an error in gender, but far more Iikely, an error in genekr and case.

(12a) * Wm stehen auf Berg.

pl / fem / mm, acc singular / masc / nom, acc, dat

(b) Wir stehen a u f h Berg.

We are standing on the rno~ntain.~

Die and Berg agree in the nominative and accusative case. Berg,

however, is singular masculine, while die occurs only with singular, ferninine

nouns. According to Holland [1994] it is likely that the student made a

mistake in gender as weU as case, because the determiner die is not a dative

determiner; yet dative is required by a state verb (stehen) with auf. In CALL

systems described by HoIland [1994], however, only an error in gender is

reported. The reason is that the noun Berg carnes the case feature of dative

which is projected up the tree, regardless of the features of the determiner.

Because dative matches the requirements of auf with stehen, no error is

flagged beyond the determiner-noun level.

Designed Intelligence: A Langaage Teacher Model 56

Page 71: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Reporting only the emr in gender, however, is pedagogicaLIy unwise.

It would UeIy cause the student to change die to de< which is still incorrect

since, although the determiner der is now the correct gender for the noun

Berg, the determiner is not infiected for the required dative case?

Example (12a) illustrates a . instance representing two underlying

emo&. m e a pmer will typically flag only one source of the error, the

teacheis feedback,,however, would be The determiner "die" iç incorrectly

inflecteed for gender and case, addressing both sources of the error. Such

feedback is not misleading and supplies the student wi th the appropriate

information to correct the error. An intelligent system should possess t h i s

pedag0gi~a.U~ important awareness.

2.2.2.2 An Unlikely Source of an Error

Unlike ambiguous errors where both sources of the error should be

reported, there are errors which may be classified as trivially ambiguous:

given fmo possible sources of an error, one is potentially so misleading o r so

unlikely that a language teacher would easily dismiss it.

(13a) * Der Gotter zürnen.

(b) Die G6tter ziifnen.

The gods are angryry8

Example (13a) containç an ambiguous error with the noun phrase.

Either the student made a mistake in number agreement between the

determiner and the noun since der is singuiar and Gotter is plural, or s/he did

not aççign the correct case to the noun phrase. Der G6tter could be a genitive

plural noun phrase; however, as a subject of the sentence, the expected case is

nominative. But as Schwind Cl9951 states "F]t is vëry unlikely that a student

should to cons*& a genitive plural, which is a 'difEcult' case, when the

Page 72: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

nominative is required." Thus a language instructor's feedback would be The

deteminer "der" is incorrectly inflected for number. However, the parçer will

not flag an error in number since der Gotter is correctly i d e c t e d for number

in the genitive. ARer the verb and the noun phrase have combined, the parser

will detect a case error and that is the error that wïü be reported.

Holland [1994] and Schwind [1995] propose a solution whereby lower

level errors are ad&essed f i r ~ t . ~ These are errors inside a noun phrase: errors 3

in gender and number between determiner and noun, for example. Schwind's

[1995] argument is that once these errors are reported t o the student, the

higher level errors, such as case concord, become more salient and can be

dedt with.

7. Rogers [1984:71 provides examples of similar ambiguity caused by adjectives. In these instances, however, it is difficult for even a human

indefmite articles and kmguage instructor to

pinpoint the exact source of an error. In the absence of additional data, theteacher wodd typically address both sources of the error. Consider examples 1-2.

(la) * Ich *de Ihnen & Urlaub empfehlen. (b) Ich -de Ihnen einen Urlaub empfehlen,

I wodd recommend a holiday for you. (2a) * offentliche Interesse (b) offentTic)ies Interesse BUT &s o f f a Interesse

public Utterest In example (la), the error in the indefinite article is due either to wrong case or wronggender; ein would be appropriate for either nominativdmasculine, neuter or accusative/neuter. Either the student incorrectly assumed that Urlaub is a neuter noun and thus chose the correct case ùiAection or the student knows that Udaub is a masculine noun but did not apply the required accusative case. In example (2a), the emor is due either to wrong gender or wrong adjective inflection. It is likely that the student incorrectly assumed that Interesse is a feminine noun since most of the nouns sufiked with -e are feminine in German. But it is also possible that the student is aware that Interesse is a neuter nom, but did not apply the correct adjective infiection for an urzpreceded adjective. The adjective chosen in example (2a) shows the correct inflection if preceded by a deh i t e article. This particular kind of amtiiguity, given in example (2211, only occurs with w r e c u adjectives since in the absence of an article the student's gender assignment to the noun is not apparent. It is therefore ambiguous whether the student applied the wronggender o r chose the wrong adjective inflectwn. In instances of modified NPs nreceded by an article, it is the article which provides the clues as to whether the student applied the wrong gender or case to the NP or whether s/he is not aware of the correct adjective paradigm.

Designed Intelligence: A Language Teacher Mode1 58

Page 73: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Addressing lower level errors first is usefd in the sense that it is a

very general d e which can be easily implemented in an ILTS. However, this

approach makes an assumption about the student's knowledge of German

grammar - that with all ambiguous errors the student is more Iikely t o have

learned case and adjective inflections than gender or number. The supposition

is unwarranted, as well as pedagogically unsound. For example, Juozulynaç

Cl9941 who analyzed an approximately 400 page corpus of German essays by

American college $tudents in second-year German courses found that

students made more case errors than errorç in number. For this reason, a

system which i d e n s e s al1 ambiguous emors as lower level errors is in many

instances likely to misrepresent the actual source of the error.

Addressing lower level enors first would also not solve ambiguou

errors where the feedback should address both sources of errors, as required

for example (12a) described previously

A M e r shortcoming of the technique used by Holland Cl9943 and

Schwind Cl9951 is that addressing lower level errors nrst is not desirable for

contingent errors. Contingent errors require a focus on higher level errors.

They will be explained in detail in Chapter 4.

The analysis described for errors in feature matching can favourably

respond to ambiguous errors due to a number of techniques which mill be

discussed in the following sections.

8, She cites m e r examples which require basically the same analysis as given for examples (12a). See Schwind 119953, p- 314. 9. Schwind [1995] refers to the technique as Case Filtering.

Designed Intelligence: A Language Teacher Model 59

Page 74: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

2.2.3 Analyzing Noun Phrases

The analysis presented here makes no distinction between lower and

higher level information. Instead, case, number, and gender agreement of the

determiner are analyzed a t the syntactic level of the verb phrase: the case

feature of a noun and the three features of its determiner, case, gender, and

number are all percolated t o the syntactic level of the verb phrase. To examine

the concept, considei the tree structure given in Figure 2.7.

NP- case

v number + gender

Det- N number, gender

Figure 2.1: Case, Number, and Gender Features

Figure 2.7 illustrates that in Gerrcan, the verb assigns grammatical

case to its çubject, that is, the noun and the determiner. The noun determines

the gender and number idiection of its determiner. As a result, there are a

number of possible errors with noun phrase constituents at distinct syntactic

levels :

Designed Intelligence: A Language Teacher Mode1

Page 75: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

1. the noun can show an error in case, and

2. the determiner can be incorrectly inflected

for case, nunber, or gender.

The three possible errors of the determiner, however, are intertwined.

Consider again the example Qted earlier by Schwind [1995].

(14a) * Der Gotter ziïmen. - - 1-

(b) Die Gotter ziirnen.

The gods are angry.

The problem with sentences like (14a) is that if the number and

gender agreement of the determiner are analyzed ùidependently of its case,

the feedback might not be accurate For instance, the gender and number

agreement of der Gotter is correct at the noun phrase level because both der

and &tter are possible constituents in the genitive plural. Only after the verb

and the subject have coinbined will an error in case be ffagged -- because the

subject requires nominative case. Yet a language instructor's feedback would

be l%e deteminer "der" is incorrectly inflected for number. The parser,

however, cannot ff ag an error in =ber because number agreement is correct

at the syntactic level of the noun phrase.

ln the technique described thus far for errors in feature matcbing, the

verb analyzes the determiner for case, number, and gender. This method,

however, requi-es that the gender and number of the noun be apparent. In

addition t o percolating all features of the determiner to the syntactic level of

the verb phrase, the phrase descriptor that records the agreement of the

determiner marks exadly which particular determiner was involved in a

given noun phrase. To achieve this, determiners use a fber-grained feature

value distinction: deremor, for example, as opposed to simply grror. This h e r

distinction d o w s the designer to fully reconstruct the error context. The - - - - - - - -

Designed Intelligence: A ILangaage Teacher Model 61

Page 76: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

following section will dismss the analysis of errors in feature rnatcbing as

applied to the sentence *Der Gotter ziinen.

2.2.3.1 Feature Percohtion

To account for ambiguous errors, the analysis of errors in feature

matching requires that a determiner be specified for a l I possible case, number,

and gender featurer For example, the lexical entry of the determiner der is

given in Figure 2.8. The features nom, dat, and g m

~hon < der >

represent case

synsem Local head

[masc corred neut dererror

nom l fem dererror plu dererror

neut dererror fem dererror

lPlu dererror ]

dat

neut dererror fem correct

Figure 2.8: Lexical Entry for der

agreement between the determiner and the verb. The three features masc,

Designed Intelligence: A Language Teacher Model 62

Page 77: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

neut. fan correlate to the three genders of a noun in the singular; the feature

& represents the agreement between a determiner and a plural noun. The

determiner specifies its value for each possible occurrence of case, number,

and gender. For instance, the value for nominative/ plural is dererror because

the nominative, plural determiner is die and not der.'' For the genitive plural,

it is correct.

- For the se)ence *Der Gotter zürnen, the n o u n Gotter determines the

gender and number of the determiner. Gotter is a plural noun. For this reason,

the nom wi l l inherit the value for the feature pLia for aU. grammatical cases

fkom the determiner, given in Figure 2.9. For clarity, the values of the

determiner have been instantiated for the phrase &r Gotter in Figure 2.10.

In addition to the three agreement constraints of a determiner, case

agreement between the noun and the verb also needs t o be analyzed. The

feature noun of Gotter, given in Figure 2.9, specifies a value for each possible

grammatical case of the n o m . For example, for nominative and genitive the

value is correct since the noun is correctly inflected for these two grammatical

cases. For dative, however, it is error- Gotter is not a dative plural noun.

For its subject, the verb zürnen, given in Figure 2.11 collects values

for the nominative case. The verb will inher i t the values for the gender and

number features mas& neut> fem, and &, respedively, from the noun. In addition, zZirnen will also-inherit the value for the feature noun representing

the case of the noun itself. Finallx the leAcal entry of ziirnen further s p e s e s

that its phrase descriptors will inherit ail values fkom the agreement featwes

of its subject.

10. Determiners can take a number of values: b e r r m , dererror. b e r r o t and m. For example, errors specified for the debite article der aU take the

value dererror, illustrated in the lexical entry of der in Figure 2.8. The error types correspond to the phonetically distinct definite articles in German.

- - - - - - - -. . . -

Designed Intelligence: A Language Teacher Mode1 63

Page 78: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

?bon < Gotter >

head ;--

n

case

I

synsem

lplu O nom noun correct

acc noun correct 1

noun correct 1

local head case

nom [plu O]

dat rPlu O]

Figure 2.9: Lexical Entry for Gotter

m e r parsing the sentence Der Gotter ziirnen, the sentence coIlected

the five phrase descriptors, given in Figure 2.12.

Note, however, that the values for the features rnasc, neut, and fem

carry the value absent since these constraints were not present in the

student's input: Gotter is a plural noun. As a result, the features masc, neut,

and fem did not get instantiated during parsing and are thus ignored by the

system. But the two phrase descriptors [mainaindause [vp-subj [plu deremorIl]

and [rnainainc1ause Evp_subj [noun -11 recorded the number, gender, and

Designed Intelligence: A Language Teacher Mode1 64

Page 79: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

phon < der Gotter >

head

n

case plu O

[no, ,,O,)

gen noun correct

Local head case ! nom [plu dererror]

acc lplu dererror 1 dat [plu dererror]

gen [*lu correct]

Figure 2.10: Lexical Entry for der G6tfer

case agreement of the'determiner and the case concord of the noun,

respectively As a result, the agreement featues of the determiner and the

noun were percolated t o the verb phrase.

While feature percolation plays a fundamental role in the treatment

of ambiguous enors, there are two additional techniques which assist in

obtaining the desired feedback. They will be discussed in the following section.

Designed Intelligence: A Language Teacher Mode1 65

Page 80: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

phon < zürnen >

head u

local rat aead

n

-

case nom

-

Figure 2.11: Lexical Entry forziirnen

1 phon < der Gotter zürnen > cat [head i

descriptor

-

main-clause

masc [~i' neut fern plu I- noun n

masc absent1 11 neut absent fem absent plu dererror noun correct 1

Figure 2.12: Phrase Descriptors forDer Gotter zümen

Designed Intelligence: A Language Teacher Mode1 66

Page 81: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

2.2.3.2 Phrase Descriptors and Feature Values

In analyzing students' input, the grammar itself does not encode the

precise source of an error. Lexical entries enumerate the possible constraint

violations, and then a phrase descriptor notes the constraint and provides the

grammatical context in which the construction occurred. As a result, the

grammar and pedagogy are kept separate. The designer can interpret the

source of the error on the basis of the error context provided by the phrase

de~c&~to r . In the fchowing example, two different sentences have generated

an error with a feminine noun but they result in distinct feedback.

(15) [main-clause lvp-subj Wern dererrorlll

(16) [main-clause [vp-dirobj [fem dererrorlJJ

In (151, the student made a mistake with a femiriine noun as the

subject. Phrase descriptor (16), however, shows an error with a feminine noun

as the direct object. In both sentences, however, the determiner der has been

used. This knowledge is apparent because determiners use a finer-grained

feature value distinction, dererroz dieerror, et al., as opposed t o simply error.

This h e r distinction allows the designer to fully reconstruct the error context

and thus encode the desired feedback. The phrase desrriptor given in (15) will

yield the feedback 2 h r e is a rnistake in gender with the deteminer "der" of the

subject because der is a possible determiner in the nominative for a masculine

but n o t a feminine noun. In contrast, in (16), the student made an error in

case; der is correct for a feminine n o m in the dative, but not the accusative as

required by a direct object. F o r such phrase descriptor, the system will

generate the feedback mere is a mistake in case with the determiner "derJ' of

the direct object.

The analysis for errors in feature matching also solves the ambiguous

error in the example *Der Gotter ziirnen provided earlier. Given the phrase

descriptor [mainclause [vp-subj [plu dererrorll], the designer can condude

Designed Intelligence: A Language Teacher Mode1

Page 82: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

that the student used der for a plural subject noun in a main clause. In such

occurrences, the feedback will dways be The determiner "der" is incorrectly

znflected for nurnber, which is the appropriate response:

F o r ambiguous errors where the student made two mistakes, both

errors should be reported. For the sentence *Wir stehen auf die Berg provided

earlier, the parser generates the phrase descnptor [main-clause (ppdat

[masc dieerrorll]. The fact that the student used the dennite article die for a

r n a s c ~ e noun in a dative prepositional phrase is apparent. In such cases,

the feedback will always be D i e determiner "die" is incorrectly infiected for

gender and case which is the desired feedback as discussed in section 2.2.2.1.

Because the approach percolates al1 features of the determiner and

the noun to the syntactic level of the verb phrase, and separates pedagogy

fkom the grammar, it is possible to encode precisely the feedback a teacher

would provide for an error in a particular context. The algorithm described

provides the information required as part of the conceptual design of phrase

descriptors. In a modular system, it becomes a simple matter to alter feedback

according to the pedagogy of the designer. The grammar does not need to

make any assumptions o n which grammatical phenornenon a student is more

likely t o Irnow, a shortcoming of HoI.Iandys Cl9941 and Schwind's [1995]

systems, nor will the sfxdent ohtain any misleadhg feedback. In addition,

such analysis does not impose any restrictions o n other classes of errors, such

as contingent errors.

In the following section, section 2.3, I will discuss the approach for

errors in word order. 1 will introduce two algorithms for creating phrase

descriptors that record verb-second and verb-ha1 position, which is of

particular interest for German. The approach chosen, however, can be applied

to other word order constraints found in other languages."

Designed Intelligence: A Language Teacher Mode1 68

Page 83: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

2.3 Errors in Linear Precedence

Existing systems anticipate errors in word order on the basis of the

native language of the student. If not anticipated, buggy rules cannot ident*

them. For example, for the sentence given in (17a), EngIiçh learners of

Germ-an wi l l tend to place the past participle in the position which resembles

English word order, that is, after the auxiliary. A buggy rule would anticipate

precisely this error by describing a stnicture where the past participle

inconectly occurs in the position derived from the English nom, successfUy

parsing the ungrammatical sentence. However, such a grammar could not

analyze the sentence given in ( c ) where the verb is in a different, but also

incorrect position.

(17a) *Ich habe & ihrn das Buch-

(b) Ich habe ihm das Buch gejg&m.

I have giuen him the book.

(c) *Ich habe ihm das Buch

By contrast, the following approach does not anticipate errors in word

order. Instead the system records where the verb occurs and thus is capable of

analyzing either enor, given in (17a) and (c).

The following sections will focus on the analyses of verb-second and

verb-ka1 position. In main clauses in German, unlike many other European

languages, the finite verb is always the second constituent of a sentence, an

damiiliar concept for most learners. Thus, this grammatical constraint

II. A word order constraint of French, for example, is the position of adjectives. Unlike English and German, the adjective follows the noun in French. The same approach as illustrated for recording verb position in Geman can be applied.

- -

Designed Intelligence: A hnguage Teacher Mode1 69

Page 84: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

presents dif5nilties. Section 2.3.1 will describe the general technique for

analyzing verb-second position.

2.3.1 General Technique for Verb-Second Position

In German, the &te verb can be followed o r preceded by the subject,

o r can even be preceded by a subordinate clause. In example (18), the subject

precedes the verb a e in example (191, due to the t h e adverbial the subject

follows the verb. In example (20) the verb is preceded by a subordinate clause.

In ail instances, the verb is the second constituent of the sentence.

(18) Er peht. He is leaoing.

(19) Morgen & er. Tomorrom he is leavirtg.

(20) Dass er geht, & ich.

%t he is Zeauing, I know.

Examples (18) - (21) reveal why the position of a finite verb in German

cannot be analyzed by simply relying on a particular grammar rule to assign

the value. At the t ime of each rule application it cannot be determined

whether the verb is indeed the second constituent of the sentence. For

example, for both the sentences given in (18) and (211, the subject er will

combine with the verb geht. Yet after the sentence has been parsed, the f i t e

verb in (18) is in the correct position, while in (21), it is in the wrong position

due to the time adverbial preceding the verb phrase.

Examples (18) - (20) have a common structure although they are

formed by different d e s . The syntactic tree, given in Figure 2.13 illustrates

that the finite verb in German always occurs as the second constituent in a

sentence regardless of t he phrasal type. Under this analysis, the grammar

Designed IntelLigence: A Language Teacher Mode1 70

Page 85: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Verb Y

Figure 2.13: Partial Structure of Finite Verb Position

spec&es a feature representing the position of a verb. All words except finite

verbs ignore the value of the position feature. The grammar takes the second

constituent of a sentence and assigns the value correct t o the position feature

of its first constituent and the value erro to the position feature of all other

constituents. The method ensures tha t the position value correct is always put

in second position while for a l l other positions, the value is error, given in

Rgure 2.14.

correct error

Figure 2.14:Feature Values for Verb-Second Position

The phrase descriptor that records verb-second position will inherit

its value fkom the position feature of the verb. The final result is a phrase

descriptor indicating whether the constraint on verb-second position has been

met. Section 2.3.2 will illustrate the concept in detail.

Designed Intelligence: A Language Teacher Modei 71

Page 86: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

2.3.2 Verb-Second Position

In analyzing a sentence, ail grarnmar rules build up a list for verb-

second position in the same way: the fkst value of the Est is assigned to the

l e b o s t constituent and the rest to the rightmost constituent. Each member

of the list represents a constituent in a sentence. Figure 2.15 (a) illustrates

the syntactic tree for verb-second position for the correct sentence Er geht.

< error, correct, ... >

NP VF' I

Er I

geht error correct

Figure 2.15: Percollation Pattern for Er geht

During parsing, the list < X, Y> is constructed where Y represents tbe

position value of the verb geht. Mer the sentence has been parsed, the

Assignqos rule instantiates the values of the list, given in Figure 2-16? the

first item of the list receives the value error, while the value for the second

constituent is correct, illustrated in Figure 2.15 (b). As a result, for the

sentence er geht, geht records the value correct for its position feature.

The grammar dso specifies that the position feature of the verb

geht is structure-shared with its appropriate phrase descriptor, given in the

partial lexical entry of the verb geht in Figure 2.17.

12. The number of terms of a list represents the longest possible sentence anticipated by the system. Each term corresponds to a constituent in a sentence. For brevity, there are only six terms given in Figure 2.16

Designed Intelligence: A Language Teacher Mode1 72

Page 87: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

cat

[ ~ y m ~ r n 10.. [cat [ h e d [il]]]]

Figure 2.16: Unary Assigngos Rule

phon c geht >

local

cat [head v]

descriptor rnaixl-c1auçe position-mainclause [finite m]] [

Figure 2.17: Phrase Descriptor for Verb-Second Position

Designed Intelligence: A Language Teacher Mode1 73

Page 88: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

The h a 1 result for the sentence Er geht is the phrase descriptor

[mainâU1clause [position-mainclause [finite correct111 indicating that the verb

geht is in the correct position.

Note, however, that this analysis will also parse sentences such as the

one given in example (22a).

(Ba) *Sie ihm hm.

(II) Sie hil& ihm.

In example (22a), the f i t e verb hilft is in the wrong position. It needs

to be the second constituent of the sentence. The analysis desrribed will parse

the sentence by the rule that combines a noun phrase with a following verb,

the Complement-Head d e , and the rule which combines a subject with a

following verb, the Subject-Head d e . The phrase descriptor which is

generated, however, will be [main-clause [position-mainclause [-nite errorll1.

To illustrate the analysis of a sentence with a verb in the wrong

position, consider the syntactic tree structures given in Figure 2.18. .

For the sentence, *Sie ihm hilff, the grammar rules wiU build up the

list c X, Y, Z > where Z represents the position value of the verb hi@. The

Assigngos d e will instantiate the values of the list as < error, correct, error,

... >, given in Figure 2.18 (b). As a result, the verb hi@ will record the value

error for its position feature. The position value is structure-shared with the

phrase descriptor that records the position value of a fhite verb. The final

result with respect to analyzing verb-second position of the sentence *Sie ihm

hile is the phrase descriptor [main-clause [position-mainclause [fuite errorll J

indicating that the constraint on verb-second position has not been met.

Designed Intelligence: A Language Teacher Mode1 74

Page 89: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

< error, correct, error, ... > S

I < correct, error > Sie

error

Y Z correct error

Figure 2.18: Percolation Pattern forSie ihm hilft

mThile the parser and grammar build up a Est for verb-second

position, verb-final position is analyzed in a diff'erent way. In the following

section, 1 will illustrate the approach for recording verb-final position.

2.3.3 General Technique for Verb-Final Position

In German, infinitives and past participIes always occur in &al

position in main clauses. Examples are given in (23) and (24), respectively

(23) Er kann pehm. He can l e a k

(24) Er hat sie getroffen.

He has met her:

Figure 2.19 shows the general structure of the position of a nodiaite

verb in German.

Designed Intelligence: A Language Teacher Mode1 75

Page 90: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Figure 2.19: Partial Structure of Nonfinite Verb Position

The nonfinite verb is in the wrong position if it occurs as the lefcmod

constituent in a phrase. If the verb is the rightmost constituent, its position is

correct only if the verb is the final constituent of the sentence. It cannot be

determined, however, whether the verb is indeed in h a 1 position until the

entire sentence has been parsed. For this reason, the grammar rules spe@ a

position enor for the leftmost constituent of a phrase. The position value for

the rightmost constituent, however, is assigned by a final rule, one which

applies only &r the sentence has been parsed. Al1 words except nonfinite

verbs ignore the value of the position feafxre. Figure 2.20 illustrates the

position values for a nonfinite verb.

error correct

Figure 2.20: Feature Values for Verb-Final Position

In addition, the phrase descriptor that records the position value will

inhent its value fkom the position feature of the nonhi te verb. The h a 1

Designed Intelligence: A ILanguage Teacher Mode1 76

Page 91: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

result is a phrase descnptor indicating whether the constraint o n verb-fkid

position has been met. The following section will provide examples of the

algorithm employed.

2.3.4 Verb-Final Position

To follow the analysis f o r verb-fmal posit ion for the correct sentence - Er kann gehen, consider the tree structure, given in Figure 2.21.

error Er A v VP

l kalm

I gehen

error Z

Figure 2.21: Percolation Pattern forEr kann gehen

Figure 2.21 reveals that the position value of the nonfinite verb gehen

is n o t assigned in the grammar d e s . Instead, the as yet uninstantiated

position value is percolated to the sentence level. After parsing the sentence,

the Assigngos ru le will assign the position value correct to the infinitive

gehen. The unary Assigngos rule is partially given in Figure 2.22.

The grarnmar dso specifies that the position feature of the nordbite

verb is stmcture-shared wi th the phrase descriptor that records verb-fkal

position, given in the lexical e n t q ofgehen in Figure 2.23.

Designed Intelligence: A Language Teacher Mode1 77

Page 92: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

-

synsem

phon c gehen >

synsem local

-

local

-

1 cat [head m] r

Figure 2.22: Unmy Assign-pos Eeule

cat [head v]

Figure 2.23: Phrase Descriptor for Verb-Final Position

The final result is the phrase descriptor [main-clause

[position-mainclause [idmitive correctlu indicating that the S t i v e is in

t h e correct position.

Designed Intelligence: A Language Teacher Mode1 78

Page 93: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Note, that the grammar will also allow sentences to be parsed with a

nonfinite verb in the wrong position. Consider the example given in (25a).

(25a) *Sie kann M e n Klaus.

(b) Sie kann Klaus Belfea.

She can help Klaus.

In example (25a), the nonfinite verb is in the wrong position. The

infinitive, helfen, needs to be in sentence-final position. The grammar

descnbed will par& the sentence by the n i le that combines a verb with a

folloming noun o r verb phrase, the Head-Complement rule, and the rule which

combines a subject with a following verb, the Subject-Head nile. The phrase

descriptor which is generated, however, will be [mainclause

To illustrate the analysis of a sentence with a nodinite verb in the

wrong position, consider the tree structure, given in Figure 2.24.

I kann

error

I heKen I

Klaus error Z

Figure 2 . a Percolation Pattern for Sie kann helfen maus

Designed Intelligence: A Language Teacher Mode1 79

Page 94: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Figure 2.24 shows that for the sentence *Sie kann helfen KZaus, the

position value of the infinitive helfen is error. The position value of helfen is

structure-shared wit the phrase descriptor that records the position value of

an infinitive. The final result is the phrase descriptor [mai-clause

[position-mainclause [infinitive error111 indicating that the infinitive is in the

wrong position.

- The two algorithms for errors in linear precedence provide a general

analysis for parsing students' ill-formed input. The potition of a finite o r

nonfinite verb need no t be anticipated. The technique for verb-final position

has also been applied to identifying verb positions in subordinate clauses in

German. PVhile the fkite verb in subordinate clauses appears clause-fùially-,

the nonfinite verb occurs second-to-last,

The main advantage of the techniques for errors in feature matching

and errors in linear precedence is that the grammar treats grammatical and

ungrammatical input identically. The grammar's main task is t o assign phrase

descriptors. As a result, the treatment of errors is very general: none of the

errors covered by these methods needs t o be anticipated, neither the parsing

nor the grammar formalism needs t o be altered and no additional rules are

required. For this reason, the system does no t s&er &om poor error coverage,

limîted user scope, o r lack of generality, problems associated with other

approaches discussed in Chapter 1.

While an argument might be made that there is an element of error

anticipation in fully s p e c m g each possible grammatical constraint in a - --

Designed Intelligence: A Language Teacher ~ ~ d e l

Page 95: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

lexical en- of a sign, there is an important difference. The systems discussed

in Chapter 1 search for a specific student error. F o r example, they search for

the infinitive to the right of the auxiliary based o n English word order. For

verb positions which have not been anticipated, however, the sentence cannot

be analyzed. In contrast, the analysis here records whether the verb is in the

correct position and if not, will dways report an error no matter where the

verb occurs in a sentence. Thus, the main referent under this approach is the

Germa language as the target language rather than specific errors baçed on

the native language of the student.

The approach applied t o errors in feature matching can also

successfully address ambiguous errors, not generally handled by existing

ILTSs with the exception of Holland [1994] and Schwind [1995J Unlike these

systems, however, the approach described does not need to make any

assumptions about which grammatical phenornenon a student is more likely

to know, nor will it produce misleading feedback. It successfully emulates a

human tutor in evaluating and responding t o a student's performance by

implementing the following h o techniques:

First, the parser percolates al1 agreement features of the determiner

and the noun to the syntactic level of the verb phrase. There is no distinction

between lower and higher level errors required since the information a t both

syntactic levels is presemed.

Second, the analysis is modular in that it separates the grammar

from pedagogy: the language grammar is applied to parsing while the

language pedagogy is implemented in the feedback. As a result, the phrase

descriptors provide the error context and the designer decides on the desired

feedback by defining which error(s) shodd be addressed.

Designed Intelligence: A Language Teacher Mode1 81

Page 96: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

While the approach presented in this dissertation provides a general

treatment of errors in feature matching and linear precedence, an

unconstrained grammar increases ambiguity Ambiguous readings will be

discussed in the follo wing chapter.

Designed Intelligence: A Language Teacher Model 82

Page 97: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

C H A P T E R 3

Multiple Sentence Readings

Introduction

A standard problem in parsing is that most sentences can be assigned

more than one syntactic structure. For example, van Noord [1991 states that

the Alvey Tools Grammar with 780 rules averages about 100 readings per

sentence on sentences ranging in length between 13 and 30 words. The

maximlfl~~ was 2,736 readings for a 29 word sentence.

Although typiczh language exercises do not contain 29 word

sentences, an uncons trained grammar generally produces more parses than

systems designed for o n l y well-formed input. Consider a system for German

which relaxes the constraint on grammatical case. Without case marking, the

parser has no way of knowing if a constituent is a subject or a verb

complement. As a result, more than one sentence analysis w d l be produced

and the errors flagged in each sentence reading can vary For instance, for the

Designed Intelligence: A Language Teacher Model 83

Page 98: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

sentence given in (1) the parser can assign at least two legitimate syntactic

structures. They are illustrated in the two tree diagrams, given in Figure 3.1.

(1) Er fragt das Mâdchen.

S Subject-Head

Det N

das ~ i i d c h e n

He is asking the girl.

Sentence Structure (a)

*S Head-Subject

Complement-Head \ NP

Specifier-Head

Det N

das Miidchen

me girl is asking him. Sentence Structure (b)

Figure 3.1: Two Analyses forEr f i a . das Miidchen - -

In sentence structure (a), er is the subject of the sentence, das

Müdchen is the direct object. The parser and grammar will not flag an error.

The sentence is correct. In the second sentence analysis, however, the subject

and the direct object are reversed: das Müdchen is the subject and er is the

direct object. But in this sentence reading, an error is flagged for the direct

object since er must be ihn in the accusative case.

The two distinct sentence analyses arise £rom two, conjoint causes:

Designed Intelligence: A Language Teacher Model 84

Page 99: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

First, due to variable word order in German, even a grammar

designed to parse only well-formed input must s p e c . d e s capable of

parsing SV0 and OVS constructions. For instance, sentence structure (b) in

Figure 3.1 is ungrammatical because the object er should be ihn as a direct

object in the accusative case. However, the Complement-Head and Head-

Subject rules are nonetheless required for the correct sentence Ihn fragt das

Madchen-

Second, the feature case which disambiguates a subject and a verb

complement in a well-formed system, has been relaxed. Since the rules for the

alternative sentence analysis are in place, the grammar and the paner wiIl

produce the t w o syntactic structures given in Figure 3.1.

NLP applications frequently incorporate techniques for s e l e d g a

preferred parse. One method is to select the fkst parse produced by the

system. For an ILTS, however, this criterion is insufEcient because the b t

parse chosen might address an unlikely error, as in the case of sentence

reading (b) of example (1). As a result, the feedback might be confusing,

misleading, o r even incorrect. A pedagogic system should approximate the

students' intentions-

Weischedel et al [1978] implemented a technique whereby the

selected parse is based on the number of errors identined. m e r all parses

have been generated, the one with the fewest number of errors is chosen.

Such a technique, unfortunately, is not suitable for a pedagogic system for a

number of reasons which will be discussed in detail in section 3.2.

The prefemed method for disambiguating multiple sentence readings

more closely follows the approach of a Ianguage teacher who considers the

likelihood of an error. The likelihood of an error takes into account the level of

Designed Intelligence: A Language Teacher Mode1 85

Page 100: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

instruction and the frequency andlor difEculty of a grammatical construction.

The paramount concern is avoiding misleading or incorrect feedback.

In analyzing student input, there are two Ends of ambiguous

sentence readings which may result in misleading and/or incorrect feedback.

For the first kind, the parser produces sentence readings which are

unlikely considerigg the level of instruction and the diEculty of a

grammatical construction. For example, German exhibits a fairly fkee word

order with respect to subject and object(s) position. However, it is doubtful

that a beginner student of German would astign OVS word order over SV0 in

a main clause: object topicalization is a rare construction, even at a more

advanced level.

The second kind contains sentence readings which are potentially so

misleading that a language instmctor would easily dismiss them. For

instance, for the sentence given in (2a), although the parser might interpret

the verb gehen as an infinitive, a language instructor would clearly iden-

the sentence as containhg an error in number agreement between the subject

and the finite verb. The ambiguity arises because the 3rd plural finite form of

a verb in German is phonologically identical to its infinitive.

(2a) *Er gehen.

(b> Er geht.

He is leaoing

The two Ends of ambiguous readings are resolved by implementing

two techniques. One uses licensing conditions t o mark sentences for the rules

which constructed them and selects the appropriate sentence reading on the

basis of the Iikelihood of a grammatical construction. The other blocks parses

f?om being generated through ad hoc techniques. Both methods assist in

Designed Intelligence: A Eanguage Teacher Model 86

Page 101: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

obtaiMg the sentence structure a student most likely intended and thus

result in the desired feedback.

The task of the licensing conditions is t o distinguish multiple

sentence readings with respect to word order. During parsing, three of the

syntactic d e s , the Subject-head, the Head-Subject-Complement, and the

Head-Subject d e each assign a distinct licensing descriptor. Any of the three

licensing conditions c m license a sentence. After parsing, the Licençing

Module prioritizes multiple sentence readings so that, in the event of a choice,

the parses are selected in a particular order. The chosen parse is passed o n to

the next module of the system for firther processing.

The order in which parses are selected c m , however, be customized.

The licensing conditions, for example, for an introductory course in German

specify that if a parse exists that contains SV0 word order, i t will be selected.

Failing to find any SV0 structure, the system will çearch for a VSO. Finallx if

no parse contains either SV0 or VSO, parses with VS word order will be

licensed.' The flexlbility of customizing the ranking of the licensing

conditions is highly desirable in a pedagogic system for the following reason.

No pedagogic system should dismiss grammatical conshctions

arbitrarily. It ideally will reflect the level of instruction and the focus of a

particular exercise. F o r example, object topicalization is not a grammatical

construction explicitly taught at the introductory level where the focus is on

1. Examples of inverted verb-subject constructions with a transitive and intransitive verb are given in (1) and (Z), respectively- Example (1) illustrates a verb-Nbjeet-cornpiement (VSO) consb-ùction, the matrix clause in example (2) shows verb-subject (VS) word order.

(1) Morgen sieh t er sie,

lbmorrow he'll see her.

(2) Dass er geht, weiss ich.

That he is leauing, I k rtow.

Designed Intelligence: A Langnage Teacher Mode1 87

Page 102: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

the grammar of more commonly used rules of German. However, if an

instructor nonetheless wanted to implement topicalization, the system should

be fleeble enough to d o w it. The ranking of the licensing conditions can be

adjusted for topicalization exercises while the default setting reflects the

more general goal of a particular level of instruction.

In contrast to the Licensing conditions, the task of the ad hoc

techniques is to block parsing of sentence readings out of hand. These are

sentence readhgs which are so misleading that a language instructor would

not even consider them. For example, ad hoc techniques disambiguate lefical

entries which are phonologically identical, illustrated earlier with the verb

gehen in example (2a). While in a well-formed grammar, phonologically

identical forms are generally mutually exclusive due to their fùlly spedied

features, in an ILTS, these very same constraints might have been relaxed.

The ad hoc techniques block such sentence readings.

In sections 3.2 and 3.3 1 wi l l illustrate how the licensing conditions

and the ad hoc techniques successfidly disambiguate multiple sentence

readings.

Licensing Module

The task of the Licensing Module is to select the sentence reading a

shident most Likely intended, even given errorful input. No computational

method can always guarantee correct selection since the system has no direct

knowledge of what the student meant to express.2 For a pedagogical system,

2. While in some cases intentions can be clarified by a query to the student, less common sentence readings do not likely warrant the disruption.

Designed Intelligence: A Language Teacher Mode1 88

Page 103: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

however, the technique c a n dosely mode1 a language instructor by taking i n t o

account the likelihood of an error.

The following analysis implements three licensing conditions, given

in (3) - (5), t o disambiguate multiple sentence readings.

. . (4) [main-clause flicense-head-subjcomp truell .--

. (5) [ma&clause Clicense-head-subj truel13

The three Iicensing descriptors are assigned in their corresponding

grammar d e s : Subject-Head, Head-Subject-Complement and Head-Subject

rule, reflecting SVO, VSO, and VS word order, respectively ARer parsing, any

given sentence will contain one of the three licençing desc~ptors. Note,

however, that these three Iicensing conditions are mutually exclusive because

each d e combines the sub~ect with the rest of the sentence.

Any one of the three licensing descriptors can license a sentence The

conditions are prioritized in the system so that, in the event of a choice, the

Licensing Module selects the parses in the order presented in (3) - (5) The

chosen parse will be passed on to the next module in the system for further

processing.

3. The concept of licensing conditions can be applied to other grammatical phenornena. For example, the parser also poduces multiple sentence readings for sentences containing

.prepositional phrases. For instance, (lbl and (Ic), below, d lu t ra te h o sentence analyses of the clause given in exampIe (la):

(la) .,-, dass er in dem Garten Liest.

(b) ..-, dass [[er [[in dem Garten] liest]]].

(c) *..-, dass [[er in] [[dem Gartenl liestll].

.-., that he is reaùing in the garden. The sentence reading given in (b) is correct: er is the subject of the phrase, [in dem Gartenl is the prepositional phrase. In sentence structure (c), however, the subject is [dem Gartenl and [er in] is treated as the prepositional phrase. Two licensing descriptors can be implemented to disambiguate the multiple sentence readings with regards to the prepositional phrase.

Designed Intelligence: A Language Teacher Model 89

Page 104: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

The following sections will provide examples of the Licensing

conditions and discuss their order. Section 3.2.1 will provide examples for the

Subject-Head and Head-Subject licensing conditions. Section 3.2.2 wiU give

examples for disambiguating Head-Subject and Head-Subject-Complement

constructions.

3.2.1 . - Subject-Head and Head-Subject Licensing . -

The licensing conditions reflect the factors a language instructor

takes in to account in evaluating a student's sentence. To examine this

concept, consider example (6):

(6) Er sieht die Frau,

He sees the wornan.

A sentence which is unambiguous in many other NLP applications

can nonetheless result in multiple sentence readings in a systern designed to

parse fi-formed input. For instance, there are two syntactic structures for the

sentence fkom example (6). They are illustrated in the tree diagrams given in

figure 3.2.

For the sentence Er sieht die Frau, er could be taken to be either the

subject or the direct object of the sentence. Assuming that the choice between

the two parses were arbi-ary, sentence structure (a) where er is the subject of

the sentence contains n o errors while sentence reading (b) would yield the

feedback mis is not the correct case for the direct object.

The introductory student probably did not intend to topicalize the

object, that is, apply OVS over SVO word order. But given the feedback

provided for (b), the learner would most likely attempt to correct die Frau,

creating an incorrect sentence - one which was correct initially4 . -- . . . - - - - -

Designed Intelligence: A Language Teacher Mode1 90

Page 105: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Det

die

He sees the woman. Sentence Structure (a)

NP

N V

1 1 Det N sieht I

die Frau

me woman sees him. Sentence Structure (b)

Figure 3.2: Two Analyses for Er sieizt die Frau

The parse selected by the Licensing Module emulates the decisions

made by a language instructor in interpreting a student's response. For an

introductory ievel, this mode1 strongly prefers the SV0 word order, given in

sentence reading (a), for the following reason: a beginner student is not being

taught about object topicalization and thus is iinlikely t o attempt the

construction. Even if the student had intended the OVS word order, in

considering the level of instruction and the f?equency and difficulty of the

grammatical constmctioh, a language instructor would assume that, more

likely, the SVO was meant.

The two licensing conditions, [mainclause Clicense-subj-head truell

and [mainclause [license-head-subj true]], disambiguate the two sentence

4. The feedback would probably be even more confusing if e r as the direct object were underlined, highlighted, or specificaUy mentioned in the feedback.

Designed Intelligence: A Language Teacher Mode1 91

Page 106: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

readings. The conditions are stored in their corresponding g r m a r d e s ,

Subject-Head and Head-Subject rule, respectively For illustration, the

Subject-Head rule with its associated licensing descriptor is partially given in

Figure 3.3.

.O cal r descriptor mainclause I

Figure 3.3: Subject-Head Rule

During parsing, the Subject-Head d e assigns the phrase descriptor

[main-clause [license-subj-head truell to sentence reading (a). For the

alternative sentence reading, the Head-Subject d e generates the phrase

descriptor [mainclause mcense-head-subject true]]. For an introductory

course, the order of the licensing conditions determines that sentence reading

(a), that is, the parse containing SV0 word order wiU be selected. Thus, the

same grarnmar c a n be used for different purposes.

The ranking of the licensing descriptors can be customized according

t o the goals of an exercise and the level of instruction. For example, a

Designed Intelligence: A Language Teacher Mode1 92

Page 107: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

particular exercise might aim at practicing object topicalization. For such an

instance, the priority ranking of the licensing descriptors can be changed such

that if a choice is given, the parse containing the head-subject licensing

descriptor wilI be selected.

While the two licensing conditions in the Subject-Head and Head-

Subject rules would be sufficient for disambiguating the two sentence

readings of example (61, the Licensing descriptor assigned in the Head-

~ubjec t -~om~lement d e is required for verb-subjed inverted constructions

in German. Examples are yesho questions, rnatrix clauses following a

subordhate clause, or main-clauses containing an adjunct as the first

eiement. The following section will pravide examples illustrating the need for

the Head-Subject-Complement licensing condition and show how it differs

Çom the Head-Subject Licensing condition.

3.2.2 Head-Subject-Complement and Head-Subject Licensing

In German, the hi te verb is always the second constituent of the

sentence. For this reason, there are a number of verb-subject inverted

constructions which also cause ambiguits In the example given in (7a), a time

adverbial precedes the verb, while the subject follows the verb.

(7a) *Heute sieht sie a

6) Heute sieht sie h.

Sentence (7a) results in two syntactic structures, given in Figure 3.4:

For sentence structure (a) in Figure 3.4 the Head-Complement rule

fïrst constructs the phrase [sieht siel. The Head-Subject rule then combines

the verb with its subject, [[sieht sie] er]]. The rule assigns the licensing

descriptor [mainâinclause mcense-head-subject truc]]. In this sentence reading

no error is flagged: sie is correctly infiected for a direct object and er for a - - -- --

Designed Intelligence: A Language Teacher Model 93

Page 108: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

sieht sie Head Complement Subject

lbday, he21 see her.

Sentence Smcture (a)

Adv S

sieht sie er

Head Subject Cornplentent

M a y , she'll see him.

Sentence Structure (II)

Figure 3.k Two Analyses for *Heu& sieht sie er

subject. For sentence structure (b) in Figure 3.4, however, the Head-Subject-

Complement rule assigns the licensing descriptor [main-clause

[licens e-head-subject-comp true]] and the feedback is The direct object is

incorrectly inflected for case since er, as a direct object should be ihn. The

correct sentence reading is that in (b). Although the parser does not flag an

error for sentence reading (a), the sentence is in fact ungrammatical. In

German, an object cannot be topicalized in inverted verb-subject

constructions. In this analysis, the ungrammatical construction is ruled out by

the ordering of the licensing conditions, given in (8) to (10).

(8) [main-clause Clicense-subj,head me13

(9) (main-clause Clicense-head-subjcomp truell

(1.0) [main-clause ficense-head-subj meIl

-- --

Designed Intelligence: A Language Teacher Mode1

Page 109: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

The order of the three licensing descriptors specifies that the parse

containing the Head-Subject-Complement descriptor, given in (9, will always

be selected over the Head-Subject licensing descriptor, given in (10); thus

eliminating the ill-formed topicalized sentence reading.

The ranking of the three licensing descriptors determines that if a

parse e i s t s that contains the SV0 word order it will be licensed. Failing t o

find any SV0 structure, the system will search for a VSO. As illustrated in the

example Heute sieht sie er, the VSO word order is also subsumed by VS order,

and must therefore be considered first to yield the desired feedback. Findy,

the Head-Subject licensing condition, given in (IO), wiU only be selected if n o

parse contains any of the other two conditions. This applies t o sentences with

intransitive verbs in inverted sentence constructions. An example is given in

(Il):

(11) Dass er geht, weiss ich.

%t he is leaving, 1 know.

The Head-Subject d e is the only rule whicb can apply to the matrix

clause combining the verb with its subject. For this reason, the descriptor

[mainclause Bcense-head-subj t rue] is required to License such sentences!

The licensing conditions mode1 a language instructor in

disambiguating multiple sentence readings in word order. The analysis

presents an alternative t o the technique first implemented by Weischedel et

5. Evea in an exercise where students practice topicalization, the Head-Subject-Complement licensing condition will always be considered over the Head-Subject condition because the former is subsumed by the latter. To allow for topicalization, the Subject-Head licensing condition would be considered last. 6. Unlike main clauses, embedded clauses do not require licensing descriptors. Under this analysis, errors in word order are reported first and parses are never arnbiguouç with respect t o verb position. Once the verb is in dause-final position, there is no ambiguity with regard t o the subject or object. IR subordinate clauses, al1 constituents occur to the left of the verb and the subject-verb d e requires a saturated complement Iist of the verb.

Designed Intelligence: A Language Teacher Mode1 95

Page 110: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

al. Cl9781 which selects the parse with the fewest number of enors. The two

are contrasted in the following section.

3.2.3 Licensing vs. Fewest Error Parse

The Zicensing approach is more suitable to feature underspecification

than a technique that is based on the fewest number of errors for a number of

reasons:

Unlike Weischedel's [1978] method, the licensing technique excludes

S e q u e n t (emphatic, poetic, etc.) sentence readings, if a choice of a more

fiequent usage is available. For example, for the sentence given in example

(12), ifsie is the subject of the sentence, then er is incorrectly infiected because

er should be ihn as a direct object in the accusative case. If sie is topicalized

and er is the subject, the sentence is correct.

(12) Sie liebt er.

She loves him or It is h r he loues.

However, a language instructor would not assume that a beginner

student intended the more poetic second reading. For this reason, the

preferred feedback would be This is not the correct case for the direct object

which corresponds to the grammatical constmction the student most likely

intended In Weischedel's Cl9781 technique, no error would be reported since,

an error-free sentence reading exists, albeit an improbable one.

To give equal weight t o syntactic structures that are grammatically

possible, but rare, would cloud feedback. Querying the student to

disambiguate sentences is one solution, but unwarranted if one alternative is

S e q u e n t . Feedback messages would double o r triple in length in order to

accommodate a tiny percentage of constmctions. A human tutor makes

assumptions about context and student sophistication. These assumptions

Designed InteLligence: A Language Teacher Model 96

Page 111: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

have been transferred as much as possible to the approach presented here. If

the student indeed intended to topicalize the direct object in the example sie

liebt er, the feedback message will not be confiising since the student is aware

of the construction. However, accepting the sentence as correct when the

student did not airn a t topicalization would leave the student with a

misconcep tion.

- In addition, the ordering of the licensing conditions can be

customized. Exercises which aim at topicalization can be implemented and

the order of the licensing conditions changed.

The licensing technique also handes multiple parses where different

sentence readings contain an equal number of errors. An example is given in

(13)-

(13) *Sie hiM er.

She helps hirn or It is h r he helps.

For the sentence given in (131, the paner produces two sentence

analyses relevant t o this discussion. In the first reading, sie is the subject and

er is incorrectly infiected for dative. If the object is topicalized, er is the subject

and sie is incorrectly i dec t ed for the indirect object. Taking into account the

leamer level, the licensing conditions wiu select the first reading. Relying on

the parse with the fewest errors does not provide a rationale for choosing one

over the other since both-readings contain one error, either with er o r sie.

Finally, a system should present a consistent sentence analyçis. The

licensing technique consistently chooses the same parse through each

iteration of the correction process. As will be discussed in Chapter 4, a system

should display one feedback message a t a tirne. Once the student responds t o

the correction, the sentence is processed for further errors. A heuristic based

on the number of errors could easily shiR sentence readings in mid-stream,

Designed Intelligence: A Language Teacher -Mode1 97

Page 112: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

leading to confusing errors messages. This applies to instances where more

than one sentence reading contain the same number of errors. To examine this

concept, consider example ( 14).

(14) *Der Groflvater sucht der Mann.

The grandfather is Zooking for the man.

The sentence given in (14) contains an error with the direct object.

There are two sentence readings: either der Grobvater o r der Mann could be

the direct object. However, in both instances the determiner has to be idected

for accusative, where der should be den. Thus, both sentence readings contain

an error in case. Assuming that the parser fiags one error at a tirne, once the

student makes the required correction, the sentence w i l l be sent for further

error checking. When the parser analyzes the sentence for the first tirne, it

might interpret der Mann as the direct object and an error in case will be

reported. If the student incorrectly changes der Mann t o d e n Mann, perhaps

assuming that the verb assigns dative as opposed t o accusative case, the

sentence is again wrong. In the next iteration process, however, the parser

might shiR interpretations and choose the alternative reading where der

Mann is the subject. This is possible because both sentence readings still

contain one error. The result is that the feedback will be confusing for the

student. The licensing technique described wiU always select the sentence

reading with SV0 word order, if a choice is given. Thus, the parser will

consistently choose the same sentence for analysis until the sentence is

correct.

The default setting of the priority ranking of the licensing technique

is aimed at the beginner level of German and assumes, like a language

instructor that a student would not attempt to use rare constructions a t that

level. For an intermediate course or even particdm exercises, the system can

Designed Intelligence: A Language Teacher Mode1 98

Page 113: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

be set to d o w for topicalization. The priority ranking of the three phrase

descnptors can be reordered to prefer OVS over SV0 word order.

I W e the licensing conditions sift possible sentence readings

according to the likelihood of an error, the ad hoc techniques block sentence

structures that arise £rom feature relaxation.

3.3 - Ad HocTechniques

If our goal is to mode1 a language instructor in evaluating students'

responses, then it is imperative that the analysis reject implausible analyses.

Whïle the licensing techniques select the most mely sentence reading, ad hoc

techniques will block undesirable sentence readings from being parsed.

Section 3.3.1 will provide an example of an ad hoc technique which has been

applied to phonologically identical finite verbs and ~ f i v e s .

3.3.1 F'ïnite Verbs vs. Tnfinitives --

Ad hoc techniques are warranted because of sentence readings arising

fiom ambiguity of lexical entries. A problem for any relaxed grarnrnar is

leecal entries that are phonologically identical, but have distinct grammatical

functions. For instance, examples (15) and (16) illustrate that the 1st person,

plural ïdect ion of a finite verb in German is identical to its infinitive. Either

form could be selected in any given parse. It does not, however, foUow that the

choice must be arbitrary. The choice of sentence readings should reflect the

student's likely intention, and avoid inaccurate o r misleadhg feedback.

(15) Heute g.eben wir.

Today we are leaoing.

Designed hteUigence: A Language Teacher Mode1 99

Page 114: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

(16) E r kann a. Nè can go.

The sentences given in (15) and (16) wiU each produce a t Ieast 2

sentence readings. For each sentence, there will be one parse withgehen as an

infinitive and one parse with gehen as a finite verb. Ifgehen is interpreted as

an infinitive, then errors in case, gender, number, and person would be

overlooked. The reason is that infinitives do not impose agreement constraints

on the& subjects and thus do not contain phrase descriptors that record these

grammatical phenomena.7 In addition, there are instances, where the

feedback might be misleading. For example, idbitives need to be in sentence-

final position in German. For the sentence Morgen gehen wir, interpreting

gehen as an infinitiue would raise an error for verb position, although the

sentence with gehen as the finite verb is correct. The multiple sentence

readings are possible because the distinctive feature of the finite vs. the

nonfinite verb has been relaxed.

There are two ways of overcoming the problem with phonologically

identical formç: either o n l y one lexical entry is used, o r each f o r m has its own

lexical entry in which case they need to be made mutudy exclusive through

implementation of an additional feature. The analysis presented here uses one

lexical entV The advantage is that no additional features are required.*

To illustrate the analysis, consider the partial lexical entry for gehen

which functions as both a f i t e verb and an infinitive, given in Figure 3.5.'

7. The agreement features are marked with the modal which appears with the infinitive. 8. In some instances, giving each form its own lexical entry can be fairly elaborate. For example, finite verbs are the head of the sentence. Thus, the only dependent relationship with respect to the head exists between the subject and the verb. An analysis which considers two distinct lencal entries for a finite verb and an infinitive either requires that each possible subject is specified for the distinctive feature of the two verb forms or that the rules which combine a subject with a finite verb block a nonfinite verb. 9. For the purpose of this discussion, the case features have been collapsed. In addition, specifications for subordinate and coordinate clauses have been omitted.

Designed Intelligence: A Language Teacher Mode1 100

Page 115: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

phon < gehen >

local

content index !

position-maincl

Figure 3.5: Lexical Entry for gehen

Figure 3.5 illustrates that the verb gehen will generate three phrase

descriptors that record agreement constraints: vp-subj, vp-num, and vpqer .

The descriptors record number, gender, and case agreement of the subject of

gehen and subject-verb agreement in number and person, respective13 In

Designed Intelligence: A Language Teacher Mode1 101

Page 116: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

addition, the phrase descriptor position-maincl will collect a position value for

gehen as a nnite verb as well as for an infinitive. Finallx to record the

agreement between an aUX1liar-y o r a modal and its verb complement, gehen

spe&es two features: and nonfinite. They are required in instances where

a student might incorrectly choose an a d i a r y with an infinitive o r a modal

with a past participle.

- To obtain, accurate feedback, the analysis requires that the

appropriate phrase descriptors for either form are generated and that they be

mapped onto the correct verb form. To achieve this, the lexical entry of the

verb gehen specifies the phrase descriptors required for both forms. The

phrase descriptors for an infinitive, however, are only reported if they are

inherited by an auxiliary or a modal. The phrase descriptors for gehen as an

infinitive d be discussed in the following section.

Phrase Descriptors for gehen as an Infinitive

To capture the constraints of an infinitive, two grammatical

phenornena have to be recorded: verb-bal position and agreement between

an auxiliary o r a modal and the n o f i t e verb form.1° An auxiliary requïres a

past participle while a modal subcategorizes for an infinitive. An example of

an error with the modal kann and a past particïple is given in (17a).

(l7a) *Er kann eeg.angen.

(b) Erkann&

He can leave.

To record verb-bal position, the lexical entry of gehen, given in

Figure 3.5, includes the phrase descriptor [main-clause [position-maincl

10. In the aetual implementation, there are two phrase descnptors for recording verb-final position: one for infinitives and one for past participles. The two phrase descriptors allow for precise feedback of the position of the two verb inflections.

Designed Intelligence: A Language Teacher Mode1 102

Page 117: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

[infinitivenoreport XJ]]? An auxïliary o r a modal will inherit the value of the

phrase descriptor. The lexical entry of the modal kann is given in Figure 3.6.

phon 4 kann > head

comps pos 1 mainel [infinitive 011

L

descriptor

descriptor

mainclause

main-clause

Figure 3.6: Lexical Entry for kann

Figure 3.6 illustrates that kann will inherit the position value of the

infùlitive fiom the descript o r [mainclause [position~rnainclause

rinfinitivenoreport a]]. Note, however, that the descriptor of the modal is

[main_clause [position-mainclause rinfinitive Xf J J While the phrase

descriptor of the modal will contribute to a student mode1 update, the phrase

descriptor [main-clause [position~mainc~ause [infinitivenoreport X]]] is

otherwise ignored by the system.

11. X stands for an as yet uninstantiated value.

Designed Intelligence: A Language Teacher Mode1 103

Page 118: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

To record the agreement between an auxiliary o r a modal and its verb

cornpiement, gehen specities two features: & and nonfinite. For the idbi t ive

gehen, the modal kann, given in Figure 3.6, spedes a phrase descriptor that

records the agreement. It will inherit its value from the feature bse of the

nonfinite verb. The rernaining phrase descriptors given in the l e ica l entry of

gehen are n o t inherited by an auxilïary o r a modal and are thus ignored for

gehen as an infinitive. This applies to the phrase descriptors that record

subject-verb agreement in case, number, gender, and person and the position

of a f i t e verb.

Phrase Descriptors for gehen as a Finite Verb

To examine the constrâints of a finite verb, the grammar requires

three phrase descriptors for recording the agreement phenornena between the

subject and the verb: vpçubj, vpger, and vp-num. The verb gehen specifies

that it will inherit the values hom its subject. In addition to agreement, verb-

second position of the n n i t e verb also needs to be analyzed. The phrase

descriptors illustrated in Figure 3.7* record the constraints of a finite verb.

position-maincl [finite ( 1

Figure 3.7: Phrase Descriptors for gehen as a Finite Verb

12. For the purpose of this discussion, the case, number, and gender features have been collapsed.

Designed Intelligence: A Language Teacher Mode1 104

Page 119: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Note, however, that the phrase descfiptor that records the position value of an

infinitive is ignored for gehen'as a finite verb. In addition, none of the phrase

descnptors of the finite verb are inherited by the modal and thus are not

recorded for gehen as an infinitive.

The ad hoc technique allows for disambiguation of multiple sentence

readings arising fkom phonologically identical lexical entries. The analysis

uses -one lexical entry and the phrase descriptors for each grammatical

phenomenon are successfully mapped ont0 the correct verb fom. As a result,

potentially misleading parses are blocked.

The technique can be applied t o further instances of phonologically

identical foms also found in other languages. For example, English presents

the same grammatical phenomenon of infinitives vs. finite forms.

Multiple parses present a challenge for any ILTS because the parser

might choose an unlikely sentence reading. As a result, the learner might

receive misleading feedback o r an error might be overlooked.

The Licensing conditions described in this chapter mark sentences for

the d e s which conçtructed them and select the appropriate interpretation on

the basiç of the likelihood of a given grammatical construction. The ranking of

the licensing conditions can be customized to reflect the focus o f a particular

exercise and/or the level of instruction,

In contrast to the licensing conditions, the ad hoc techniques block

sentence interpretations that are potentially so misleading that a language

teacher would easily dismiss them. The technique described c m be applied to

further instances of phonologically identical forms.

Designed Intelligence: A Language Teacher Mode1 105

Page 120: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Both methods described in this chapter disambiguate multiple

sentence readings by emulating the language instructor in providing

intelligent feedback to students' input.

Designed Intelligence: A Language Teacher Mode1 106

Page 121: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

CHAPTER 4 Pedagogically- Informed Feedback

4.1 Introduction

Analyzing student input in an ILTS requires linguistic and pedagogic

knowledge. The task of the grammar and the parser is to generate phrase

descriptors which provide the linguistic analysis of students' input. The

phrase descriptors, however, need to be processed M e r to reflect principles

of language learning. The Analysis Module, the Student Model, and the

Filtering Module, respectively, constitute the pedagogic components carrying

out the analysis described.

The Licensing Module, discussed in Chapter 3, chooses the desired

parse. AU phrase descriptors associated with the selected parse are then sent

t o the Analysis Module, and the information obtained is subsequently passed

to the Student Model and Filtering Module. Figure 4.1 illustrates the relation

between the Analysis Module, the Student Model, and the Filtering ~odule. '

Designed Intelligence: A Language Teacher Mode1 107

Page 122: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

AU Phrase Descriptors ikom the Selected Parse

Analvsis Module

returns student model updates and produces instructional feedbac i of increasing abstraction

Student Model

keeps learner model updates and decides o n the student level 1

Filterine Module

decides o n the order of instructional feedback

Error-contingent Feedback Suited to Learner Expertise

Figure 4.1: Pedagogic Moddes

The Analysis Module takes all incoming phrase descriptors and

generates possible instructional feedback at different levels of granularity

comelated with student expertise. In ILTSs, granularity is particularly

important t o fkaming responses t o learners' errors. Inexperienced students

1. Of these three modules, only the Analysis Module is language-dependent but to a trivial extent: the feedback messages in the current implementation happen to be in English.

Designed intelligence: A Language Teacher Mode1 108

Page 123: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

require detailed instruction while experienced students benefit best from

higher level reminders and explanations naReau % Vockell 19891.

For instance, in example (la) the student made an error with the

determiner einen of the prepositional phrase. Von is a dative pïeposition and

UrLaub is a masculine noun. The correct article is einern.

(la) *Sie traurnt von eineq Urlaub. -

fi) Sie t&umt von ~inem Urlaub.

She is dreaming of a uacation.

In a typical student-teacher interaction, feedback depends on the

students' previous performance history. For instance, for the error in example

(la), a leamer who generdy has mastered the concept of dative assigning

prepositions might merely receive a hint indicating that a mistake occurred

within the prepositional phrase. For the learner who generally knows the

grammatical r u l e but s t i l l needs practice in its application, the feedback is

less general: in addition t o location, the teacher might point o u t the m e of the

eaor (case). For the novice learner, the feedback would be as speciiic as

possible. In addition to the location and type of the error, information as to the

exact source of the error, that is, the precise grammatical rule that has been

violated (dative case), would be provided.

Note, however, that even the most specific report aimed at the novice

leamer refkains fkom revealing the correct answer. For the error in example

(la), the beginner student is s t i l l required t o decide on the correct Sec t ion of

a masculine indefinite article in the dative case. The pedagogical principle

underlying this design is guided discovery learning. According t o Elsom-Cook

[1988], guided discovery takes the student dong a continuum fiom heavily

structured, tutor-directed learning t o where the tutor plays less and less of a

role. Applied to feedback, the pedagogy scales messages on a continuum fkom

least-to-most specific guiding the student towards the correct answer.

Designed Intelligence: A Language Teacher Mode1 109

Page 124: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Burton and Brown LI9821 state that hinting a t the source of an error

supports the development of students' self-regdation, a prerequisite for

guided discovery learning. A system can lead the learner towards the correct

answer rather than simply supplying it, thus assisting the students in finding

the source of an error themselves. The ability to discover the source of an

enor, however, strongly correlates with the expertise of the student: the more

exp- the leamer the less explicit feedback is necessary Fischer & Mandl

19881.

G r a n u l e captures this pedagogicdy intuitive concept: the

Analysis Module generates three categories of instructional feedback

corresponding to three learning levels: expert, intermediate, and novice. The

responses are s d e d fkom least-to-most specific, respectively Provided with

three categories of feedback by the Analysis Module, the Student Model

selects the error response suited to the student's expertise according to a

record of the student's previous performance history o n a particular

grammatical phenornenon.

The Student Model tracks 79 ~~ar constraints, corresponding t o

the grammatical concepts being monitored in an introductory course for

~ e r m a n . ~ For each of the grammar constraints, the Student Model keeps a

counter which, at any given instance in the evaluation process, falls in the

range of one of the three learner levels. If a grammatical constraint has been

met, the counter is decremented. If the constraint has not been met, the

counter is incremented and a feedback message suited to the performance

level of the leamer is selected.

2. The grammar constraints have been chosen according to Wie geht's, the course book commonly used in an introductory course for German fsee Sevin, Sevin, & Bean 199U.

Designed Intelligence: A Language Teacher Mode1 110

Page 125: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

A . example of a grarnmar constraint is pp-dat which records the

student's performance on dative assigning prepositions. Each student is

initially assessed as an intermediate, which has been chosen as a reasonable

default l e ~ e l . ~ A leamer who violates the constraint on dative prepositions

will, a t first, obtain the feedback message for the intermediate. If the student

commits the same error in subsequent exercises, s h e will soon be assessed a

novice. At this point, the Student Model will select the more detailed feedback

message suited to ihe beginner. However, each time the student applies the

grammatical constraint correctly, the Student Model records the success- &?ter

demonstrating ~roficiency, the student will again be assessed as intermediate,

or, even expert. Maintaining a large number of grammatical constraints

allows for a very detailed portrait of an individual student's language

cornpetence over a wide-range of grammatical phenomena.

The Filtering Module decides on the order in which instructional

feedback is displayed. Feedback is provided for one error at a time to avoid

overloading the student with extensive error reports. According to van der

Linden [1993], displaying more than one feedback message at a time makes

the correction process too complex for the student.

The ordering of the instructional feedback is decided by an Error

Priori@ Queue which ranks the 79 grammar constraints maintained by the

Student Model. The ordering of the g r m a r constraints and ultimately

feedback is, however, adjustable. Depending o n the pedagogy of a particular

language instructor, they can be reordered t o reflect the focus of a par t idar

exercise. In addition, errors which are not relevant to a particular exercise

3. The intention of the system is primarily to augment but not to replace classroom instruction. The system is therefore a practice tool for the student who previously has been exposed to a particular grammatical concept of Gennan- It is thus safe to assume that the learner will not be a complete beginner nor an expert on the grammatical phenomena practiced.

Designed Intelligence: A Language Teacher Model 111

Page 126: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

need not be reported at dl, although the error itself' still is recorded in the

Student Model.

The ordering of the grammar constraints also accommodates

contingent errors, a special class of multiple errors, and thus avoids

misleading or redundant feedback. F o r instance, in example (2a) the student

made a mistake with the prepositional phrase. The verb denken

subcategorizes for the preposition an which requires the accusative pronoun

dich, while von is a dative preposition requiring the dative pronoun dir. If we

igaore the dependency of the errors with the preposition and the pronoun, the

feedback t o the student would be "von" is the wrongpreposition and This is the

wrong case of the pronoun "dich". However, the pronoun dich is not incorrect if

the student changes the preposition von to an because an requires the

accusative pronoun dich. Depending on the order of the feedback, the student

might end up chmghg von to an and dich t o dir and wind up utterly

confused. S h e might ultimately fd to fmd the correct case of the pronoun

because it had been flagged as an error in the original sentence.

(2a) * Ich denke y~a. dkh-

dat prep. acc pronom

(b) Ich denke dich.

acc prep. acc pronoun

I am thinking of you.

The error in the pronoun is correctly flagged by the system f?om a

purely logical point of view. However, fiom a pedagogical perspective reporting

the eaor in the pronoun dich is redundant and possibly misleading. In such

instances, only the error in the preposition von is reported although the error

in the pronoun is recorded in the Student Model. Thus, the system, while

silently recording dl errors, does not necessarily comment on all of them.

Designed Intelligence: A Language Teacher Mode1 112

Page 127: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

The £bal result of the three modules is error-contingent feedback

suited to students' expertise. The following sections will discuss the Analysis

Module, the Student Model, and the Filtering Module in detail.

4.2 The Analysis Module

The function of the Analysis Module is to take all incoming phrase

descriptors as input and retum student model updates and potential error

feedback at dinerent levels of granulari@ for each phrase descriptor.

Granularity has been previously applied to an Intelligent Tutoring System for

LISP programming [McCalla & Greer 1994, Greer & McCalla 19891. In their

system SCENT, Greer and McCalla implemented granularity to "recognize

the strategies novice students employ when they solve simple r e m i v e LISP

progranunhg problemç."4 in t h i s analysis, however, granuiarity is used in

&aming responses t o learners' errors: inexperienced students obtain detailed

instruction while experienced students receive higher level reminders and

explanations. For example, consider the ungrammatical sentence in (3a).

(3a) *Der Mann dankt dem Eiwu. -.

(b) Der Mann dankt der Baia,

The man thanks the woman.

In example (3a), the student has provided the wrong determiner for

the indirect object. For the error dem Frau, the system generates feedback of

increasing abstraction that the instruction system can use when interacting

with the student. The level of the learner, either expert, intermediate, o r

novice according t o the current state of the Student Model, determines the

4. Greer & McCaIla C19891, p. 478-

Designed Intelligence: A Language Teacher Model

Page 128: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

particular feedback displayed. The responses, given in (4a) - (c) correspond to

the three leamer levels for the error in example (3a), respectively:

(4a) There is a mistake with the indirect object,

(b) There is a rnistake in gender with the indirect object.

(c) This is not the correct article for the indirect object. The noun is feminine.

For the expert, the feedback is most general, providing a hint t o where - Z

in the sentence the error occurred (indirect object). For the intermediate

leamer, the feedback is more detailed, providuig additional information on the

type of error (gender). For the beginner, the feedback is the most pretise. It not

only pinpoints the location and type of the error but also refers to the exact

source of the error (feminine noun).

Figure 4.2 displays the partial Granulârity Hierarchy for conçtraints

in feature matchingg5 The Granularity Hierarchy is a representation of the

instructions given to the student conelated with the grammatical constraintç

monitored by a phrase descriptor. Each term in a phrase descriptor

corresponds t o a level in the Granularity Hierarchy For example, for the

indirect object of the sentence *Der Mann dankt d e n R a y given in (3a) on p.

113, the grammar and the parser will generate the phrase descriptor

[main-clause [vp-indirobj [fem errer]]]. The top node of the Hierarchy

specitîes in which kind of clause the error occurred. The phrase descriptor

indicates that a mistake- was made in a min-clause. The next level in the

Granulari@ Hierarchy Lists possible errors in each clause type. As indicated

by the phrase descriptor, the mistahe refers t o the indirect object. A n even

her-grained constraint speafication is found in the next lower level of the

Granularity Hierarchy For instance, an indirect object can be incorrectly

5. The hierarch~ here is simplified for the purpose of illustration. The lowest nodes in the hierarchy split into further nodes as illustrated with gender.

Designed Intelligence: A Language Teacher Mudel 114

Page 129: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

idected for either case, number, o r gender. The phrase descriptor specifies

that an error in gender occurred, çpecifically with a femirzine noun which

corresponds to the lowest level in the Granularity Hierarchy

Main Clause Coordinate Clause Subordinate Clause

Decreasing . S pecificity

Semantic Constraints

Case Gender Number

Preposition Semantic Constraints

Verb Inflection ModaVlnfhitive

Figure 4.2: Granularity Hierarchy for Constraints in Feature Matching

Each node in the Granularity Hierarchy corresponds to a level of

speciûcity of a feedback message. Granularity works dong one dimension,

namely, abstraction: travelling downward through the Hierarchy, nodes lower

in the Hierarchy will have assotiated with them messages of increasing

specificity. Generally, the more experienced the student the coarser-grained

the message and the higher the node.

The Analysis Module is implemented in DATR [Evans and Gazdar

19901, a language designed for pattern-matching and representing multiple

inheritance. Nodes in DATR are represented by the name of the node followed

Designed Intelligence: A Language Teacher Model 115

Page 130: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

by paths and their values. For example, the node given in Figure 4.3

corresponds to the phrase descriptor that records the position of a &te verb

in a subordkate dause.

('possubclfin' '3' 'true' T h e verb in the subordinate clause is not in the correct position.'

'possubclfm' '2' 'true' T h e h i t e verb in the subordinate clause is not in the correct - .

' posit-ion.'

'possubclh' '1' 'true' T h e finite verb in the subordinate clause is not in the correct

position, I t has to be the last element of the sentence.')

<sub-clause position-subclause fmite correct> = ('possubclfin' '1' 'false' "

'possubclfin' '1' 'false' "

'possubclfin' '1' 'false' ")

esub-clause positionçubcIause finite absent> -- (").

Figure 43: DA= Code Listing for a Finite Verb in a Subordinate Clause

The paths in a node defbition represent descriptions of grammatical

conçtraints monitored by phrase descriptors. The matching algorithm of

DATR selects the longest path that matches leR to right. Each path in a node

is associated with atoms on the right of '=='. For example, if there has been an

error in word order in a subordinate clause, the parse will contain the phrase

descriptor [sub-clause [position~subclause [ f i t e error]]] . This will match the

path csub-clause position-subclause fmite er ron which s p e d e s four atoms

for each of three groups. Each group represents a learning level. The three

learning levels considered are: expert, intermediate, and novice6, given in

Figure 4.3.

6. IFiner distinctions can be made by looking at the actual error count either for the purposes of evaluation and remediation; however, three levels are sufficient ta distinguish among the feedback messages encoded.

Designed Intelligence: A Language Teacher Model 116

Page 131: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

The f k s t atom in each group specifies the grammar constraint as

described by the incoming phrase descriptor. For example, the grammar

constraint possubclfi represents finite verb position in a subordinate clause.

The second atom specifies a value that is decremented or incremented

depending on whether a grammatical constraint has been met o r not,

respectively

- - The Boole& values, as the third atom, indicate whether it is an

increment o r decrement: true for increment, false for decrement.

Finally, the fourth atom specifies a feedback message. For example,

for the path csub-clause position-subclause finite erron provided earlier, the

specificity of each response corresponds to the grammar constraints in the

Granularity Hierarchy which fiames responses to constraints on verb

position, given in Figure 4.4.

The three instructional responses associated with each node

correspond to the three leamer levels. The feedback for the beginner student

reflects the lowest node in the Granularity Hierarchy, while for the

intermediate and expert student the message will refer t o nodes higher in the

hierarchy. In a typical

more profitient in the

becomes less specinc.

student-hacher interaction, as the student becomes

use of a grammatical construction, error feedback

If there has been no error in the student's input, the phrase descriptor

is [sub-clause [position~subclause [finite correct]]] and no message is

associated with the name of the grammar constraint. However, the first three

atoms, the grammar constraint, the decrement, and the Boolean value are

7. A phrase descriptor that contains the value &=.t is ignored by the system. Thus, the list t o the right of '==' is empty.

Designed Intelligence: A Language Teacher Mode1 117

Page 132: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

1 Verb Position 1 Decreasing Specificity Main Clause

Subordinate Clause Coordinate Clause

A

- - - - - - - - - - - -

Finite

Infinitive

Past Participl e

Figure 4.4: GranulariSr Hierarchy for Constraints in Linear Precedence

still spe&ed, given in Figure 4.3 on p. 116. They contribute to a Student

Model update to record the success.

Finallx if the phrase descriptor is [sub-clause [position-subclauçe

[finite absent]]], then the grmat ica l phenomenon was not present in the

student input. As a result, the phrase descriptor is ignored by the system,

indicated by an empty list in Figure 4.3 o n p. 116.

The &al output of the Analysis Module is a student mode1 update

and a list of possible instructional responses h m a coarse to fine grain size

for each incoming phrase descriptor. The more expert the student with a

particular grammatical construction, the more general the feedback.

Designed IntelIigence: A Language Teacher Model 118

Page 133: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

The modular design of the system also allows the encoding of

instructional feedback suited to the pedagogy of a particular teacher. F o r

instance, an instructor who preferred to avoid the term finite verb in beginner

messages could substitute the simpler verbS8

The list of alternative instructional responses fkom the Analysis

Module has to be processed further before being displayed t o the student. The

Student Model, which picks a response suited to the level of the student, will

be dismssed in the following section.

4.3 The Student Mode1

Individualization of the learning process is one of the features of a

student-teacher interaction that distingtishes it &om gross mainçtreaming of

students characteristic of workbooks. Students l e m at their o w n pace and

often, work for their own purposes. Leamers also vary with respect to prior

language experience, aptitude, andior learning styles and strategies [Odord

19951. According to the Individual DSerences Theory as described by Oxford

C19951, if learners learn diaerently then they Likely benefit fiom

individualized instruction. An ILTS can adapt itself to different learner

n e e d ~ . ~ The requiremerit, however, is that the system incorporates and

8. There is an argument to be made for softerllng technical terms in the intermediate and novice categories. Rather than implementing a pseudo-vocabulary, however, a system can underline or highlight the error and provide hot links to information and examples on the linguistic terminolog' The advantage of using preuse linguistic terminology for al1 learner levels might result in students not only learning the grammatical concepts but the proper terminology a t the same time. 9. In its strict definition, ILTSs which do not implement a student mode1 are ICALL systems rather than ILTSs. However, in this thesis al1 systems which make use of Natural Language Processing are referred to as ILTSs.

Designed Intelligence: A Language Teacher Mode1 119

Page 134: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

updates knowledge about the learner. Student modelling provides the key to

individualized knowledge-based instruction McCalla & Greer 19921.

Student modelling has not been a strong focus of paner-based ILTSs,

likely because the challenging taçk of representing the domain knowledge in

ILTSs is s t i l l largely incomplete molland & Kaplan 1995b1.'~ If the grammar

is not accurate and complete, even a precise student model cannot

compensate. For inStance, HoIIand Cl9941 states that a system which doeç not

flag ambiguous and contingent errors accurately will obscure a student

model. A further possible reason is that many significant dif3erences in

language learning swles (situational, aural, visual, etc.) are precluded fkom

consideration in text-based ILTSs.

ILTSs which do have a student model primarily concentrate on

subject matter performance. M o d e b g students' surface errors assists in

individualizing the language learning process and 3 s sufncient to model the

student to the level of detail necessary for the teaching decisions we are able

t o make."ll The technique aids in teaching the required skills and

remediating the student.12

10. According to Holland and Kaplan C1995bJ there are two trends with student modelling in systems for language learning. The North American focus lies on the NLP module, the domain knowledge. Due to the c o m p l e x i ~ of the NLP component, these systems implement student models only to the extent. of analyzîng surface errors. In contrast, Eumpean systems concentrate on more sophisticated student models by hypothesizing causes of enors (see Chanier et al., 119921). However, they "opt for narrow NLP bandwidthn. Holland & Kaplan 11995b1, p. 364, Il. Elsom-Cook [19933, p. 238. 12. The rnodelling techniques employed in language learning are distinct fiom those applied to procedural tasks. In procedural tasks, the modelhg techniques presume that the student can learn the skill in terms of a number of formal steps. Yet although languages are rule- govemed systems and we can represent linguistic ability in terms of formal step-by-step niles, we do not produce language by consciously following them [Bailin 19901. As a result, the modelling techniques in language learning primarily diagnose the sources of errors rather than model strategies which the student used in solving a particular problem.

Designed Intelligence: A Language Teacher Model 120

Page 135: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

McCaIla Cl9921 makes a distinction between implicit and explicit

student m o d e h g which is particularly useful in classi&ng the student

models in ILTSs.

An implicit student model is static, in the sense that the Student

Mode1 is reflected in the design decisions inherent to the system and derived

fkom a designeis point of view. For instance, in an ILTS the native language

of the leamer can be encoded as a bug model and ultimately used to diagnose

errors.

In contrast, an explicit student model is dynamic. It is a

representation of the leamer which is used to drive instructional decisions.

For W S s , for instance, the student model can assist in guiding the student

through remedid exercises o r i t can adjust instructional feedback suited to

the level of the learner. In either case, the decisions are based on the previous

performance history of the learner. The following discussion will provide

examples of KTSs which have implemented implicit and explicit student

models.

4.3.1 Implicit Student Models

Implicit student modelling has been applied to ILTSs to diagnose

errors. For example, in-Catt & Hirst's Cl9901 system Scripsi the native

language of the student represents the learner model. It is used to model the

leamer's interlanguage. With regard to student modelling, the pitfall of such

an implementation is that it is a static conception. The system3s view of the

learner cannot change across interactions with the system. It has no impact

o n instructional decisions and provides only a gross individualization of the

learning process when ideallx a student rnodel is dynamic Bol t et al. 19941.

Designedl Intelligence: A Language Teacher Mode1 l.21

Page 136: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

In a more hdividualized example, Bull Cl9941 developed a system

that teaches clitic pronoun placement in European Portuguese. The student

model is based on the system's and the student's belief measures, language

learning strategies, and language awareness.

The system's belief rneasure is comprised of the proportion of

incorrect/correct uses of the rule; the students provide the data for the

studiea.tk belief measuse, being required to state t h e 5 confidence in their

answer when entering sentences. Learners also iden* their preferred

leaming strategies when using the program. Accordhg to Bull [1994],

language awareness is achieved by dowing the student access to all

information held in the system. None of the information, however, is used to

drive the instructional process. In addition, a number of studies have s h o m

that students tend to not take advantage of the option tu access additional

information. For example, Cobb & Stevens Cl9963 found that in their reading

program learners' use of self-accessible help was villtually non-existent, in

spite of their previously having tried it in a practice session, and also having

doubled the success rate as compared to either a no help or dictionary help

option in the practice session.

4.3.2 Explicit Student Models

In developing & explicit student model one typically starts by

making sorne initial assumptions based on pretests or stereoSpical

postdations about the leamer. For example, initially every student could be

asçessed as an intermediate. During the instructional process, the student

model adjusts to student's behaviour moving to a novice o r expert profile, as

appropriate. This technique is used in explicit student models t o make

instructiond decisions.

Designed Intelligence: A Language Teacher Model 122

Page 137: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Explicit student modelling has been used in a number of ILTSs,

primarily in the form of tracking. Tracking can be as simple as calculating

percentages of correct answers o r more sophisticatedly, i d e n e g particular

errors which occurred in the student's input. The information is then used to

alter the instructional process, either in the form of further language tasks or

feedback,

- - Explicit student modelling is found in the system The Fawlty A h d e

Tutor [Kurup, Greer & McCalla 19923 which teaches correct article use in

English. The system presents the student with scenarios whereby the student

must select the correct article form and the appropriate rule. The tutor keeps

an error count and selects the scenarios on the basis of the performance of the

student; thus the path through the program is individualized by altering the

instructional process according to prior performance of the student.

Bail.in [1988,1990] in his system Verbcon /Diagnosis also employs the

tracking method. Diagnosis provides practice in using English verb forms in

written texts. All verbs are presented in their infinitid form challenging the

student to provide the appropriate verb form. The system tracks the most -

fkequent error occurrence and the context in which the error occurred. The

information is used to provide informative feedback based o n contrasting

correct and ungrammatical uses of tenses. In addition, Diagnosis suggests

exercises to help with the remediation process. .

The student mode1 under the analysis of this dissertation is based on

students' prior performance and it ultimately has two main functions. The

first is to select instructional feedback suited to learner expertise and second

t o use the student's performance for assessrnent and remediation.13 The

following section will discuss the technique employed.

Designed Intelligence: A Language Teacher Model 123

Page 138: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

4.3.3 Feedback Suited to Learner Expertise

To select instructional feedback suited t o the level of the learner, the

Student Model keeps track of 79 grammar constraints. These grammar

constraints correspond t o the grammatical concepts to be monitored in an

introductory course for Geman. The grammar conskaints are split among

the three clause types, main, subordhate, and coordinate, each containing 25

nodes. In addition, there are two gramrnar constraints monitoring noun

phases practiced in isolation, one for verb-initial position in main-clauses,

and one for the entire sentence.

The Student Mode1 assumes three types of leamers: the novice, the

intermediate, and the expert. Each student level is represented by a range of

vdues14:

novice: 20 < X I 30

intermediate: 10 I X < 20

expert: O I X < 10

hitially, the leamer is asseçsed with the value 15 for each grammar

constraint, representing the mean score of the intermediate learner. The

values are used by the Student Model to decide on the specincity of the

instructional feedback being displayed. The intermediate learner has been

chosen as a reasonable default. While the messages might be in i t idy too

overspecified for the exp.ert and too underspecified for the novice, they will

quickly adjust t o the actual leamer level.

individualize the mode1 fkom the outset.

13. The system does not yet contain any pre-defined

Pre-testing could be used to

exercises, but in an extended version error counts held in the ~ G d e n t Model be available to the l e k e r . They form the basiç for branching decisions and remediation recommendations. 14. The ranges chosen for each student level roughly correspond to the average size of a grammar exercise unit. Since the default decrement for the error count is 1, suc ces^ completion of 10-15 exercises will record a change in student level,

Designed Intelligence: A Lanpage Teacher Mode1 124

Page 139: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

For each learner level, the Student Mode1 receives four atoms from

the Analyçis Module, described previously: the name of a gramrnar

constraint, an inmernent/decrement, a Boolean value and in case of the latter

being true, a feedback message.

The counter of the grammar constraint determines which of the three

leamer responses is selected. If the number is 20 or greater, the system

displays the message suited for the beginner. If it is less than 10, the system

treats the learner as expert, and any number in between points to an

intermediate. The counter is bounded a t O and 30-

Once the Student Model has selected the feedback message to be

diçplayed to the learner, the counter is incremented by the incoming weighted

constant. For the expert, the increment is generally 3, for the intermediate 2,

and for the novice 1. The consequence of this sequence is that in case of errors,

the student wiU switch quickly from expert to intermediate, and somewhat

less quickly &om intermediate to novice. As a result, if a student is not an

expert with a given grammatical construction aRer all, the feedback wiIl

quickly become more informative fkom the expert to the intermediate level.

The transition 60m the intermediate t o the begimer level is slower since the

feedback for the former already points at a specific error type. For instance, for

the grammar constraint possubclfin, given on page 116 the feedback for the

intermediate level specifies The finite verb in the subordinate clause is not in

the correct psi t ion. l5

15. Ultimately, i t might be desirable to incorporate knowledge of the student's past performance- For example, at the end of each practice session the current error count for a particular grammar constra.int could be averaged with a historical count representing the last N sessions. By considering a histox-ical error count in deterrnining student level, momentary lapses in performance would be balanced by previous performance.

Designed Intelligence: A Language Teacher Model 125

Page 140: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

In the case of a correct response, as indicated by the Boolean value

false received fkom the Analysis module, the constant of the grarnmar

constraint is çubtracted from the counter. The decrement for al1 graRîmar

constraints is 1. Thus, the transition f%om novice to intermediate to expert is

gradual. The result of assigning a small decrement at all learner levels is that

any student has to apply a correct construction many times before the

feedback becomes sparse.

The Student Model presented has a number of advantages. I t takes

into account students' past performance, and by adjusting the value to be

incremented o r decremented, it is adaptable to a particular grammatical

constraint in an exercise or the pedagogy of a particular instructor. For

example, a language instructor might rate some errors more saLient than

others in a given exercise. In such an instance, the increment/decrement of

some grammar conçtraints can be tuned to change their sensitivity.

Its main strength, however, lies in the fact that a single e r r o f i

sentence will not drasticdy change the overall assessment of the student.

The phrase descriptors collect errors that indicate precisely the grammatical

context of the mistake. This enables the Analysis Module to create grammar

constraints which reflect verg specSc grammatical concepts that are

maintained in the Student Model. The consequence is that a student can be at

a diaerent level for any given grammar constraint reflecting the performance

of each particular g r m a r skill. This is desirable in a language teaching

environment because as a student progresses through a language course a

single measure is not sunicient t o capture the lmowledge of the learner and t o

distinguish among leamers. The Student Model described allows the student

to be geared toward enor-contingent and individualized remediation.

Designed Intelligence: A Language Teacher Model 126

Page 141: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

The Student Model can also be used for assessrnent and remediatioc.

A branching program can be implemented where students' tasks are

determined by their prior performance. The grammar constraints can be

weighted for a partinilar set of exercises, so as to be especidy sensitive to

salient errors. In an authorhg program, these settings could be adjusted by

the language instructor.

- Finallx the data obtained can be used for further research in

improving the overall performance of the system, and might prove usefüi in

providing objective measures of student performance.

The final step in analyzing students' input is handled by the Filtering

Module. The task here is to accommodate multiple errors. Multiple errors

have been largely overlooked in ILTSs. From a pedagogical perspective,

however, instructional feedback messages need to be prioritized by the system

and displayed one at a time to the student to avoid multiple error reports and

redundant and/or misleading feedback in the case of contingent enors. The

following discussion will focus on this task

4.4 The Filtering Module

While it is desirable to construct a system capable of detecting and

accurately explaining all errors, it does not follow that the system should

display each and every error detected. In the absence of an error filtering

mechanism, the sheer amount of feedback would overwhelm the student. For

example, in evaluating her own system Schwind [1990a] reports that

"[s]ometimes, however, the explanations were too long, especially when

students accumulated errors."16 In a language learning exercise, a student

Desigaed Intelligence: A Language Teacher Model 127

Page 142: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

might make more than one error. However, a language instructor typically

skips krelevant errors, and discusses the remaining ones one a t a t h e .

Example (5a) illustrates an example of multiple errors.

(5a) * Heute meine &dem haben mit & Ball.

(b) Heute haben meine Kinder mit dem Ball &- Today my children were playing with the ball.

In example (5a) the student made the following five errors:

1. word order: the finite verb haben needs to be in second position

2. word order: the nonfinite verb gespielt needs to be in final position 3. spelling error with the verb spielen 4. wrong plural inflection for the subject =rider 5. wrong case for the dative determiner dem

From a pedagogical and also motivational point of view, a system

should no t ovemhelm a student with instructional feedback referring to more

than one error at a time. Schwindys [1990a] solution to this problem is that

rnuItiple errors should be avoided from the outset. She suggests that sentence

construction exercises should focus on s p e s c grammatical phenornena such

as prepositions o r verb cases [see also Kenning & Kenning 19901.

While Schwind's approach is probably inherent in many ILTSS~~,

lUniting the teaching domain is only a partial solution. Even a basic sentence

in German, as illustrated in example (5a), requires a number of rules and

knowledge about the case system, prepositions, word order, etc.l8

Little research has been done in Computer-Assisted Language

Learning regarding the volume of feedback for different kinds of learners at

16. Schwind (1990a1, p. 577. 17. ILTSs concentrate on sublanguages to also achieve a higher degree of accuracy. [Levin & Evans 19953. 18. Holland [1994], in her system BRIDGE, displays only one error at a time, and permits instructors to divide errors into primary, which are automatically displayed, and secondary, which are displayed only at the student's request.

Designed Intelligence: A Language Teacher Mode1 128

Page 143: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

different stages in their language development. However, van der Linden

[1993] found that "feedback: in order to be consulted, has to be concise and

precise. Long feedback (exceeding three k e s ) is not read and for that reason

not use^."^^ She M e r states that displaying more than one feedback

response at a time makes the correction process t oo complex for the student

[van der Linden 19931.

- Van der Lirgien's [1993] study makes three final recommendations:

r. feedback needs to be accurate in order to be of any use to the student,

2. displayhg more than one error message at a time is not veIy usefid because a t some point they probably will not be read, and

3. explanations for a particular error should also be kept short.

With regard to feedback display, van der Linden's Cl9931

recommendations require a system submodule to sift al1 incoming errors. The

errors have to be reported one at a time and the emor explanations should be

brief This provides the student with enough information to correct the error,

but not an ovemhelming amount, and yet records detailed information within

the student mode1 for

The analyçis

ranks student errors

assessrnent and remediation.

described implements an Error Priority Queue which

so as to display a single feedback message in case of

multiple constraint violations. The ranking of student errors in the Error

Priority Queue is, however, flexible: the grammar constrâints can be reordered

t o reflect the desired emphasis of a particular exercise. In addition, a language

instructor might choose not t o report some eaors. In such an instance, some

grammar constraints will display no feedback message at all, although the

19. Van der Linden C19931, p. 65.

Designed Intelligence: A Language Teacher Mode1 l29

Page 144: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

error will s t i l l be recorded in the Student Model. The following section wiU

discuss the Error PrioriQ Queue.

4.4.1 The Error Priority Queue

The Student Model maintains grammar constraints and selects

instructional feedback suited to leamers' expertise. In case of multiple errors,

the Error Priority, Queue determines the order in which instructional

feedback messages are displayed t o the leamer. It ranks instructional

feedback with respect to

1. the importance of an error within a given sentence, and

2. the dependency of errors of syntactically lower and higher constituents.

The Error Priority Queue for the grammar constraints of a main

clause is partially given in (61.~' The names of the grammar constraints

generated and rnaintained by the Analysis Module and Student Model,

respectively are given in parentheses.

(6) Error Prior i ty Queue

1. Word Order in a Main Clause

1. position of a finite verb in a main-clause (posmainclfin) 2. position of a nonfinite verb in a main-clause (posmainclnonfm) 3. position of a finite verb in initial position (posmainclinitial)

direct Obiects in a Main Clause

1. case of the noun of the indirect object (indirobjnounmaincl) 2. case, nuniber, and gender of the determiner of the indirect object

(in dirobjmaincl)

1. case agreement of conjoined nouns of indirect objects (indirobjconjnounmaincl)

20. A complete Es t of the grammar constraints implemented in the system is provided in the Appendix,

Designed Intelligence: A Language Teacher Mode1 130

Page 145: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

2. case, number, and gender of conjoined determiners of indirect objects (indirobjconjmaincl)

N- Preriositional Phrases in a Main Clause

1. choice of preposition (prepmaincl)

2. case of the noun of a prepositional phrase in the accusative (ppaccnounmaincl)

3. case, number, and gender o f the determiner of a prepositional phrase in the accusative (pp-accmaincl)

4. case of the noun of a prepositional phrase in the dative (pp-datnounmaincl)

5. case, number, and gender of the determiner of a prepositional phrase in the dative (pp-datmaincl)

Each grammar corstraint, given in the Error Prioriw Queue in (6)

correlates to a node in the Granularity Hierarchy, provided earlier. The

grammm constraints are grouped according to grammatical phenornena. For

example, the group preuositional phrases in a main clause contains all

constraints monitored w i t h prepositional phrases and it corresponds to the

node prep. phrase in the Granularity Hierarchy in Figure 4.2 on p. 115. Each

member of a group in the Error Priority Queue refers to a node lower in the

Granularity Hïerarchy: the grammar constzaint choice of preposition, for

example, correlates with the node preposition in the Granularity Hierarchy.

The groups in the Error Priority Queue are sorted according to the

importance of an error within a sentence and the members within the group

are sorted according t o the dependency of errors of syntactically Iower and

higher constituents. If the student made multiple enors, the system ranks

instructional feedback messages according t o the order specined and displays

them one a t a time.

The Enor Priority Queue shown in (6) reflects the default setting for

the importance of an error in a given exercise. For example, grammar

constraints 1. t o 3. of the group Word Order in a Main Clause refer t o errors in

linear precedence. In the default setting, they are reported first since word

Designed InteUigence: A Language Teacher Mode1 131

Page 146: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

order is one of the fundamental concepts of a language and thus likely to have

high priority in most exercises.

The ordering of the groups of grammar constraints can, however, be

altered to reflect the pedagogy of a particular language instnictor. For

example, an instructor might want t o centre exercises around dative case

assignment. In such an instance, the grammar constraints can be reordered so

that . érrors of indirect objects are reported first. In addition, a language

instructor might choose t o suppress some errors, so as not to distract the

student from the main task. These would be errors which are not relevant to a

particular exercise. Suppressing certain errors, does not affect their

contribution to the Student Model, on the rationale that behind-the-scenes

information should be as detailed as possible.

The E n o r Priority Queue also takes into account contingent errors, a

special class of multiple errors. Contingent errors are due to a dependency

between syntactically higher and lower constituents and can result in

redundant or even misleading feedback. Contingent errors require that the

syntacticdy higher constituent be reported e s t . The following section will

discuss the process in detail.

4.4.2 Contingent Errors

The E r r o r Priority Queue also rads g r m m constraints according

to the position of the error constituents in the syntactic tree. For example,

grammar constraints, 1. - 5. of group on p. 131, refer t o prepositions and

their noun complements. Here it is especially important to report the error on

the preposition before an error with its noun complement to avoid redundant

and misleading feedback due t o the contingency between these errors. To

Designed Intelligence: A Language Teacher Mode1 132

Page 147: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

illustrate the problem with contingent errors, consider example (7a) as noted

by Schwind [1990a].

(?a) * Ich warte fii~ dk

acc prep. dat pronoun

(b) Ich warte &-

acc prep. acc pronom

I am waifing for you.

In example'(7a), the student made two errors: Für is an acnisative

preposition which requires the accusative pronoun dich. In addition, warten

subcategorizes for auf and not for fùr. The lurking cause here is likely

language interference. The English verb to wait subcategorizes for for, thus

the misuse of für. For the two errors flagged in example (7a) the student would

receive the feedback This is the wrong preposition and This is the wrong case

of the pronoun. However, the case of the pronoun depends solely o n the

preposition. In example (Ta) the pronoun is in the incorrect case for either

preposition, für and ouf both of which take the accusative. However, feedback

on contingent errors can even mislead the student. This applies to instances

where two prepositions require different cases, as given in example (8a):

(8a) * Sie denkt a ibn.

dat prep. acc pronoun

Co) Sie denkt rn ihn.

acc prep. acc pronoun

She is thinking of him.

Example (8a) illustrates that denken subcategorizes for the accusative

preposition an, while von is a dative preposition requiring the dative pronoun

ihm. As in example (7a), the feedback to the student would be mis is the

wrong preposition and mis is the wrong case of the pronoun. However, the

pronom ihn is not incorrect if the student changes the preposition von t o an

because an requires the accusative pronoun ihn. Depending on the order of

Designed Intelligence: A Language Teacher Model 133

Page 148: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input
Page 149: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Figure 4.5: Contingent Errors

An example of a system which reports the redundant emor of a

syntactically lower constituent i s desaibed by Schwind C1990aJ who reports

how students evaluated such ETS:

".-., fiequently a wrong preposition for a verb was chosen and then the wrong case for that preposition. Our system then explained both errors but some students felt tha t they did not need to h o w the comect case of a preposition which they were not going to use anyway. Clearly i t would be possible not to check the agreement of the case of a noun phrase and the preposition when the preposition is already wrong, but this means that the student is left with a misconception about a grammar rule which has been detected by the s y ~ t e r n . ~ ~ ~

As Schwind states, the student should n o t be left with misconception

in the case of contingent errors. However, as human tutors, we do not

inundate students with error explanations, particularly when an error is not

immediately pertinent. A human tutor might make a mental note to monitor

22. Schwind [l990al, p. 577.

Designed Intelligence: A Language Teacher Mode1 135

Page 150: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

the lower-level agreement construct, but would focus on the more important

error.

Only a few scholars molland 1994, Schwind 1990a, 19951 have

addressed the problems of contingent errors. Holland [1994] suggests several

improvements to her own system:

",.- omitting error classification entirely and showing students only the locations of errors in a sentence, which are then stored in sturtents' performance records. However, this means losing the relevant information about enor classification in the clear cases. "... surpressing the outermost flags in a nested series and making it obligatory to address the innermost flags. The rationale is that having students first correct lower ievel errors; typically the less serious gender disagreements gives higher level grammatical problems a chance to present themselves. This strategy makes sense for a flag series in which the outermost flag is perfectly redundant, however, this strategy runs the risk of misrepresenting student error for certain other flag configurations."23

The suggestions made by Holland Cl9941 are unsatisfactory.

The h t solution is a step backward in the development of ILTSs. For

example, Nagata [1991, 1995, 19961 studied the efficacy of simply u n d e r k g

errors as opposed to providing the student with detailed feedback about the

nature of the error and she found that the latter is indeed more effective and

more appreciated by students.

The strategy of addressing syntactically lower level (innermost) errors

&st, given in 2., is strongly motivated by the need t o account for ambiguous

errors, Houand Cl994 and Schwind [1990a, 19951, as discussed in Chapter 2.

However, as Holland Cl9941 states 'the strategy runs the risk of

misrepresenting student enors for certain other flag which

23. Holland [1994], p. 249. For the purpose of discussion, the suggestions made by Holland Cl9943 have been reordered. Also, Holland [1994] list a third suggestion which, however, is very general and provides no recommendation for a c t d implementation.

- - - - -

Designed Intelligence: A Language Teacher Model 136

Page 151: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

are precisely those contingent errors illustrated in examples (7a) - For

these, i t is the syntactically higher level (outermost) errors, such as the choice

of a wrong preposition which need to be reported first. The problem with both

Holland's Cl9941 and Schwind's [1990a, 19951 approaches is that they attempt

to handle ambiguous and contingent errors with basically the same technique.

The assumption is that one can process either higher level errors o r lower

level errors first. Yet as with ambiguous errors, any system which cannot deal

witL contingent errors effectively runs the risk of producing misleading

feedback, and M e r misrepresents the student's knowledge of the

grammatical structures involved. As a result, the Student Model will not be

accurate.

The analysis described can successfully address ambiguous and

contingent errors because a separate technique for each class of errors is

employed. Ambiguous errors are handled by feature percolation and fher-

grained feature values speafied in phrase descriptors, as discussed in Chapter

2. In contrast, contingent errors are resolved by the Filtering Module. The

Fïitering Module s a s incoming errors in such a way that the çyatactically

lower level error is not reported to the student. This is achieved by the Enor

Priority Queue which specifies that errors of syntactically higher constituents

are reported first. For example, if the student chooses an incorrect preposition,

the error will be reported before an error with its noun complement. The error

with the noun complement WU, however, be recorded in the Student Model.

Once the error with the preposition has been corrected successfiill~ the error,

if still present, with the noun complement will be addressed. The analysis

24. Holland 119941, p. 249. 25. The same shortcoming is found in Schwind's [1995] system. As discussed in Chapter 2, Schwind [1995] also addresses lower-level errors first to handle ambiguous errors.

Designed Intelligence: A Language Teacher Mode1 137

Page 152: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

ensures that the system does not misrepresent the studenfs howledge, but

at the same time, provides the student with relevant and accurate feedback.

The grammar constraints of the E r o r Priori@ Queue, can always be

reordered to reflect the pedagogy of a particular language instructor. However,

the ranking of the grammar constraints referring to contingent errors, as with

grammatical constraints 1. - 5., which correspond to the choice of a particular

prepqsition and its noun complement, constitute a fked subgroup of grammar

constraints. While the whole group c a n be assigned priority over other

grammar constraints, the order of individual grammar constraints within the

group is fked to ensure that contingent errors are addressed in a

pedagogically sound wax

The Analysis Module, the Student Model, and the Filtering Module

are essential elements of the approach presented in this dissertation. They

incorporate pedagogical and psychological aspects of language learning

essential to an ILTS that models a language instructor in evaluating and

responding to student input.

The strength of .the Analysis Module lies in its ability t o generate

instructional feedback of different granulari* ~ranularity is implemented

unidimensionally ammging gritlllIIlar constraints and instructional feedback

fkom a coarse t o fine grain size

The Student Model classifies students according t o three performance

levels: the novice, the intermediate, and the expert. 79 g r m a r constraints

are maintained by the model, reflecting the grammatical constraints to be

Desïgned Intelligence: A Language Teacher Mode1 138

Page 153: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

met in an introductory course for Germa. A student can be at different levels

for any given grammar constraint reflecting performance on each particular

grammar skill. As a consequence, a single enor will not drastically change

the assessment of the student. The information stored can be used for

tadoring instructional feedback suited to the level of the leamer and also for

assessment and remediation.

- The Filtering Module is responsible for handling multiple errors. The

approach to reporting multiple errors is pedagogically motivated, also taking

into account contingent errors. The system displays one response at a time

and the ranking of feedback responses is established by an Eaor Priori@

Queue. The Error Priori@ Queue is, however, flexible and can be easily

adjusted to reflect the focus of a particular grammatical constraint in a given

exercise. The &al result of the analysis is a single error message tailored to

the level of the leamer and pedagogical considerations.

Designed Intelligence: A Language Teacher Model 139

Page 154: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

CHAPTER 5 Conclusion

The design of an Intelligent Language Tutoring System presented in

this dissertation emulates a language instrudor by evaluating and

responding to student performance in foreign language exercises. The

techniques employed generate error-contingent feedback suited t o student

expertise,

The approach presented solves several extant problems in ILTSs.

First, errors in feature matching and linear precedence are treated in a very

general way: the main referent under this design is the target language rather

than spedic errors based on the native language of the student. As a result,

the system does not s d e r fIom limited error coverage, a narrow

preconception of the user, o r lack of generalia all problems associated with

anticipating errors based on the source language of the student. Second, by

Designed InteUigence: A Language Teacher Mode1 140

Page 155: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

emphasizing a pedagogicdy-informed, student-centered approach, the

techniques successfully address ambiguous readings, contingent and

ambiguous errors. Third, the andysis is modular making it adaptable to

specific student needs andior to reflect the pedagogy of a particula. language

instructor. Furthemore, although Gerrnan is used by way of example

throughout this dissertation, the rnodular design facilitates the task of

developing ILTSs for other languages. Language-dependencies, that is, the

domain knowledge o f the system are codîned t o the gramnar, which exists

within an othermise predominantly language-independent shell.

The approach presented consists of five components in total: The

Domain Knowledge, the Licensing Module, the Analysis Module, the Student

Model, and the Filtering Module. The five components closely correspond to

the steps a language instructor takes in evaluating a student's response to

foreign laquage exercis es.

The Domain Knowledge provides the linguistic analysis. It consists of

a Natural Language parser which analyzes student input according t u the

lmowledge o f a language encoded in the grammar. There are two main classes

of errors considered: errors in feature matching and errors in Iinear

precedence. Errors in feature matching are addressed in the lexical en- for

a linguistic sign; errors in word order are handled ui the grammar rules.

The main task of the Domain Knowledge is to assign phrase

descriptors. A phrase descriptor is a mode1 of a particular grammatical

phenomenon such as case assignment o r number agreement. A phrase

descriptor records whether or not the grammatical phenomenon is present in

the input and correctly or incorrectly formed. Ultimately, a phrase descriptor

speafies the contextual occurrence of a grammatical constraint, without,

however, deciding on the exact source of an error. - -

Designed Intelligence: A Language Teacher Model

Page 156: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

The remnining components add pedagogical value to the overall

design. Due to structural ambiguity found in language, a parser produces

more than one parse for many sentences. The Licensing Module chooses fkom

among parses by taking into account factors a language inshc tor would

consider in evaluating a student's response. In deterrnining the most likely

sentence reading, a language instructor considers the level of instruction, the

frequency of the grammatical construction, and thus the likelihood of the

error.

The Analysis Module incorporates a language instructor's knowledge

in pinpointing the precise source of an error. The Analysis Module takes

phrase descriptors as ioput and generates sets of instructional feedback of

increasing abstraction correlated with students' expertise. The Analysis

Module implements the pedagogical principle of guided discovery learnïng.

The more expert the learner, the Iess explicit the feedback required to lead

the student toward the correct answer. The modular design of the analysis,

however, makes it possible to customize feedback for a particdar instmctor o r

course.

The level of the learner, either expert, intermediate, o r novice

according to the m e n t state of the Student Model, determines the particular

feedback displayed. The Student Model keeps track of the learner history and

provides learner model updates. It records mastery of grammatical structures

as well as structures with which the leamer has problems.

The Student Model maintains a record of each grammatical

constraint to be monitored in an introductory course for German. As a resdt,

a student c m be at a meren t level for any given grammar constraint

reflecting performance of each particular grammar SU. The Student Model -- - - - -- -- - - -

Designed Intelligence: A Language Teacher Model

Page 157: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

allows for individualization of the leaming process by correlating feedback

with student expertise.

The Filtering Module decides on the order in which instructional

feedback is dispiayed. Feedback is provided for one error at a time t o avoiding

overloading the student with extensive error reports. The ordering of the

instructional feedback is decided by an Error Priori@ Queue which ranks

gramkir constraink freely with respect to the importance of an error in a

given exercise and more restrictively t o handle dependencies between

syntactically higher and lower consti~ents. The ordering of the grammar

constraints and ultimately feedback is, however, adjustable. Depending on

the pedagogy of a particular language instructor, the grammar constraints

can be reordered t o reflect the focus of a particular exercise. In addition,

errors which are not relevant t o a particular exercise need not be reported at

all, although the error itself stiu is recorded in the Student Model. Thus the

system, while silently recording all errors, does not necessarily comment on

aU of them.

The design of an ILTS outlined in this dissertation solves many

computational and pedagogical problems found in earlier ILTSs. Significant

progress has been made toward the goal of an intelligent, student-centered

tutor, but, inevitably, further work remains. For instance, while HPSG is well-

suited for the a l g o r i t h m described, there are, however, errors which pose

problems for KPSG as a head-driven grammar formalism. These concern

errors in omission and insertion. In addition, a number of pedagogical

research questions suggest themselves. The system and underlying design

will undoubtedly mature over time, benefiting from the experience gathered

through practical use. The following section will discuss further research.

Designed Intelligence: A Language Teacher Model 143

Page 158: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Further Research

Errors in omission and insertion cause problems for any ILTS

because these errors are highly unpredictable. Although missing and exfxa

constituents Gan, t o some extent, be traced to language interference, these

errorS can certainly occur due to causes other than native language

intederence. For instance, a learner might even be aware of the grammar

d e s of the target language but due to a lack of Sping skills, coxentration,

o r fatigue fails to demonstrate that understanding. For this reason, any

system which anticipates these Ends of errors on the basis of the native

language of the student sufFers fiom a narrow preconception of the user,

limited enor coverage, o r lack of generality

The approach taken in this dissertation is consistent with the overall

HPSG philosophy in that it maintains a s m d set of d e s . For this reason, the

grammar strictly eschews buggy rules that trg, for example, t o p m e a phrase

missing a word. Yet HPSG does not immediatdy offer any tools for handluig

errors in omission and insertion

briefly discuss missing and extra

of example.

in a general way. In the following 1

adjuncts, complements, and heads, by

Adjuncts are optional in that they contribute to the semantic content

of the whole phrase. Homever, missing or extra adjuncts do n o t cause a

violation of the subcategorization list of a head. In contrast, a tighter

dependency holds between complements and their heads. Heads

subcategorize for their complernents and thus a head determines the number

and kinds of complements it requires to be saturated. For errors in omission,

Designed Intelligence: A Language Teacher Mode1 144

Page 159: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

for example, the three distinct types of constituents present varying degrees

of di f f icu l ty for an ILTS.'

From a computational point of view, the least problem is posed by

adjuncts. A missing adjunct d l not cause a parse t o fail. However, from a

pedagogical perspective adjuncts still need to be addressed. For instance, in

example (la) the student has omitted an article within the adjunct, the

prepasitional phrase. The sentence will parse; however, without additional

techniques the error will not be reported- In a chart-based parsing technique

such as the one described in this dissertation, a solution might be found in

examining the chart.

(la) *Er Lest ein Buch in Garten.

(b) Er l ied ein Buch -- He is reuding a book in the garaen-

In contrast, missing complements cause a violation of the

subcategorization List of a head. For instance, in example (2a) the direct object

is missing. Unlike English, the verb sehen is always transitive in German.

(2a) *Ich sehe.

(b) Ich sehe den Stem.

1 see t h star.

For missing complements a technique might make use of the

subcategorization principle of HPSG [see Hagen 19941. If, a t the end of a

parse, the verb hzs not-found its complement(s) as specined in its lexical

entm then the verb remainç unsaturated and unsaturated words signal an

error response. Tracing verb complements presents some challenge but, at

1, There are a few errors of omission and insertion which are covered by the technique used for errors in feature matching. For example, an error with a missing auxiliary does not present a problem for the analysis in this dissertation. The sentence will parse with a nonfinite verb. A phrase descriptor indicating a missing finite verb, however, will be generated.

Designed Intelligence: A Language Teacher Mode1 145

Page 160: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

least, a record in the chart is present.

The problem becomes intractable if the head of the sentence, the verb

itselfis missing. In a head-driven approach, there is no record in the chart if

the head is missing and thus these errors require at least some extension to

the theory of HPsG?

Problems simiiar to those found with miçsing constituents are

apparent with errors in insertion. Consider example (3a):

(3a) *Er ist & Englander.

(b) Er ist Englander.

He is an Englishrnan.

Example (3a) illustrates the classic example of English speakers

incorrectly inçerting the indefinite article before nationalities in German.

However, as in the case with rnissing constituents, errors in insertion are

highly unpredictable. A sfxdent might even repeat a word in a sentence due

to, for example, a lack of concentration.

For extra constituents, HPSG also cannot offer an immediate tool for

a general solution because HPSG maintains a small set of general schemata

that do not refer to specific categories of their constituents. Thus, an answer

outside the grammar formalism has t o be found. For example, scanning the

chart for singleton words and an algorithm which drops the word in question

would be considered,

From a pedagogical perspective, a research question of further

interest concerns student rnodelling. For example, due to the implementation

2. The fact that the phrase descriptors are also head-driven is a secondary problem. If the grammar is capable of recognizing that the verb is missing and instructional feedback cm be provided to the leamer, then as soon a s the sentence contains a verb, the phrase descriptors of the verb complements can be collected.

Designed Intelligence: A Laquage Teacher Mode1 146

Page 161: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

of phrase descriptors in the overd modular design, the system is capable of

referring not only t o the source but also, in the case of native language

interference, to the cause of e r r ~ r s . ~ A study with actual language learnerç

might reveal whether feedback which addresses both the source and the

cause of an error would enhance the language learning process for different

learners at different stages in their language development.

- A b-roader study might address the role of an KTS in a given language

learning environment. For example, parser-based ILTSs inherently focus o n

grammaticd form rather than content. However, advances in multimedia

development have enabled ILTSs to d o w for a more communicative focus

without, however, neglecting form. Given this potential, empirical research

could reveal whether ETSs can form the basis of a language learning activity

rather than seroing as a second- best, that is, a mere mechanical adjunct to

dassroom grammar instruction.

3. For errors in linear precedence, for instance, a her-grained error distinction similar to that applied to ambiguous enors can be implemenbd. For example, the position value of a verb could reAect the error due to native language interference.

- - -- - -

Designed Intelligence: A Language Teacher Model 147

Page 162: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Appendix Grammar Constraints

position of a finite verb in a main clause (posrnainclfin)

position of a nonfinite verb in a main clause (posmainclnonfin)

position of a finite verb in initial position (posinitid)

inflection of the verb in a main clause (vformmaincl)

inflection of a nonfinite verb in a main clause (compverbmainclj

choice of auxiliary vs. modal in a main clause (auxmainci)

number agreement in a main clause (vp-nurnmaincl)

person agreement in a main clause (vp-permaincl)

case of the noun of a noun phrase (npnoun)

case, number, and gender of the determiner of a noun phrase (np)

case of the noun as the subject in a main clause (subjnouamaincl)

case, number, and gender of the determiner as a subject in a main clause (subjrnaincl)

case agreement of conjoined nouns as subjects in a main clause (subjconjnounmaincl)

case, number, and gender of conjoined determiners a s a subject in a main clause (subjconjmaincl)

- - case of the noun as a direct object in a main clause (dirobjnounmaincl)

case, number, and gender of the determiner as a direct objectin a main clause (dirobjmaincl)

case agreement of conjoined nouns as direct objects in a main clause (dirobjconjnounmaincl)

case, number, and gender of conjoined determiners a s a direct object in a main clause (dirobjconjrnaincl)

case of the noun as an indirect object in a main clause (indirobjnounmaincl)

case, number, and gender of the determiner as an indirect object in a main clause (indirobj maincl)

Designed Intelligence: A Language Teacher Mode1

Page 163: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

case agreement of conjoined nouns as indirect objects in a main clause (indirobjconjnounmaincl)

case, number, and gender of conjoined detenniners as an indirect object in a main clause (indirobjconjmaincl)

choice of preposition in a main clause (preprnaincl)

case of the noun in an accusative prepositional phrase in a main clause (pp-accnounmaincl)

case, number, and gender of the determiner in an accusative prepositional phrase in a main clause (pp-a&maincl)

case of the noun in a dative prepositional phrase in a main clause (pp-datnounmaincl)

case, number, and gender of the determiner in a dative prepositional phrase in a main clause (p pdatmain cl)

semantic constraints in a subordinate clause (semmaincl)

position of a finite verb in a subordinate clause (possubclfin)

position of a nonfinite verb in a subordinate clause (possubclnonfïn)

S e c t i o n of the verb in a subordinate clause (vformsubcl)

inflection of a nodinite verb in a subordinate clause (compverbsubcl)

choice of auxiliary vs. modal in a subordinate clause (auxsubcl)

number agreement in a subordinate clause (vp-numsubcI)

person agreement in a subordinate clause (vpgersubcl)

case of the noun as the subject in a subordinate clause (subjnounsubc2)

case, number, and gender of the determiner as a subject in a subordinate clause

(subjsubcl j

case agreement of conjoined nouns as subjects in a subordinate clause (subjconjnounsubcl)

case, number, and gender of conjoined determiners as a subject in a subordinate clause (subjconjsubcl)

case of the noun as a direct object in a subordinate clause (dirobjnounsubcl)

case, number, and gender of the determiner as a direct object in a subordinate clause (dirobjsubcl)

Designed Intelligence: A Language Teacher Mode1 149

Page 164: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

42. case agreement of conjoined noms as direct objects in a subordinate clause (dirobjconjnounsubcl)

case, number, and gender of conjoined determiners as a direct object in a subordinate clause (dirobjconjsubcl)

case of the noun as an indirect object in a subordinate clause (indirobjnounsubcl)

case, number, and gender of the determiner as an indirect object in a subordinate clause Cindirobjsubcl)

case agreement of conjoined nouns as indirect objects in a subordinate cIause (indirobjconjnoun%bcl)

case, number, and gender of conjoined determiners as an indirect object i n a subordinate clause (indirobjconjsubcl)

choice of preposition in a subordinate clause (prepsubcl)

case of the noun in an accusative prepositional phrase in a subordinate clause (ppaccnounsubcl)

case, number, and gender of the determiner in an accusative prepositional phrase in a subordinate clause (ppaccsubcl)

case of the noun in a dative prepositional p h a s e ui a subordinate clause (pp-datnounsubcl)

case, number, and gender of the determiner in a dative prepositional phrase in a subordinate clause (pp-datsubcl)

semantic constraints in a subordinate clause (semsubcl)

position of a finite verb in a coordinate clause (poscoordclfin)

position of a nonfinite verb in a coordinate clause (poscoordclnonfin)

inflection of the verb in a coordinate clause (vformcoordcl)

inflection of a nonfinite verb in a coordinate clause (compverbcoordcl)

choice of auxiliary vs. modal in a coordinate clause (auxcoordcl)

number agreement in a coordinate clause (vp-numcoordcl)

person agreement in a coordinate clause (vp-percoordcl)

case of the noun a s the subject in a coordinate clause (subjnouncoordcl)

case, number, and gender of the determiner as a subject in a coordinate clause (subjcoordcl)

Designed Intelligence: A Language Teacher Mode1 150

Page 165: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

case agreement of conjoined nouns a s subjects in a coordinate clause !subjconjnounsubcl)

case, number, and gender of conjoined determiners as a subject in a coordinate clause

(subjconjsubcl)

case of the noun as a direct object in a coordinate clause (dirobjnouncoordcl)

case, number, and gender of the determiner as a direct object in a coordinate clause

(dirobjcoordcl)

case agreement of .conjoined n o m s as direct objects in a coordinate clause

(dirobjconjnounsubcl)

case, number, and gender of conjoined determiners a s a direct object in a coordinate

clause (dirobjconjsubcl)

case of the noun as an indirect object in a coordinate clause (indirobjnouncoordcl)

case, number, and gender o f the determiner as an indirect object in a coordinate clause

(indirobjcoordcl)

case agreement of conjoined nouns a s indirect objects in a coordinate clause

(indirobjconjnounsubcl)

case, number, and gender o f conjoined determiners as an indirect object in a coordinate

clause (indirobjconjsubcl)

choice of preposition in a coordinate clause (prepcoordcl)

case of the noun in an accusative prepositional phrase in a coordinate clause

(pp-accnouncoordcl)

case, number, and gender o f the determiner in an accusative prepositional phrase in a

coordinate clause (pp-acccoordcl)

case of the noun in a dative prepositional phrase in a coordinate clause (pp-datnoun)

case, number, and gender of the detenniner in a dative prepositional phrase in a

coordinate clause (pp-datcoordcl)

semantic constraints in a coordinate clause (semcoordcl)

an error in any clause type (sentence)

Designed Intelligence: A Language Teacher Mode1 151

Page 166: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Ahmad, K et al. (1985). Computers, Language Learning and Language Teaching. Cambridge University Press.

Alessi, S. M., and Trolip S., R. (1985). Cornputer-Based Instruction: Methods and Development. Englewood CliEs, New Jersey: Prentice-Hall, Inc.

Badin, A. (1995). "AI and Language Learniag." Intelligent Language Trrtors: Theory Shaping Technology. Holland, M. V., Kaplan, J.D., Sama, M.R., eds. Mahwah, New Jersey: Lawrence Erlbaum Associates, Inc.: 327-343.

Bailin, A. (1990). "Skills in Context and Student Modeling." CALICO Journal, 8: 7-22.

Bailin, A-, and Thomson, P. (1988). "The Use of Natural Language Processing in Computer-As sisted Language Instruction." Cornputers and the Hurnanities, 22: 99-110.

Barker, P., and Yeates, H. (1985). Introducing Ccrnputer-Assisted Learning. London: Prentice-Hall International.

Bolc, L., ed. (1978). Natural Language Communication with Computers. Springer Verlag.

Brown, D. H. (1994). Principles of L a n s a g e Learning and Teaching. Englewood CWs, New Jersey: Prentice Hall, Inc.

Bull, S. (1994). " ~ t u d e k Modelling for Second Language Acquisition." Computers and Education, 23 (1-2): 13-20.

Burns, H. et al, eds. (1991). Intelligent î k tor ing Systems: Euolutions i n Design. Hillsdale, New Jersey: Lawrence Erlbaum Assoc., Inc.

Burton, R. K., and Brown, J. S. (1982). Investigation of Computer Coaching for Znformal Learning Activities." Intelligent TWoring Systems. Sleeman, D. and Brown, J. S. eds. Academic Press.: 79-98.

Designed kitelligence: A Langaage Teacher Mode1 152

Page 167: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Carpenter, B., and P ~ M , Gerald. (1994). The Attribute Logic Engine: User's Guide, Version 2.0. Computational Linguistics Program, Carnegie Mellon University, Pittsburgh.

Catt, M., and Hirst, G. (1990). "An Intelligent CALI System for Grammatical Error Diagnosis." Computer Assisted Language Learning, 3: 3-27.

Chanier, T., Pengelly, M., Twidale, M., and Self, J. (1992). "Conceptual Modeling . in Error Analysis in Computer-Assisted Language Learning Systgms." Intelligent ntoring Systems for Foreign Language Leaming. The Bridge to International Communication. Swartz, M. L. and Yazdani, M., eds. Nato AS1 Series F: Computer and Systems Sciences. Vol: 80. Springer Verlag.

Chen, L., and Kurtz, B. L. (1989). 'XTRA-TE: Using Natural Language Processing Software to Develop an ITS for Language Learning." Proceedings of the 4th International Conference on Artifxial Intelligence and Education: 54-70,

Chomsky, N. (1986). Barriers. Cambridge, MA: MIT Press.

Chomsky, N. (1981). Lectures on Government and Binding. Dordrecht: Foris.

Cobb, T., and Stevens, V. (1996). "A Principled Consideration of Computers and Reading in a Second Language." The Power of C . . Pennington, M. C . ed. Houston: Athelstan Publications: 115-137.

Colmerauer, A. (1978). Metamorphosis Grammars. Natural Language Communication with Computers. Bolc, L., ed. Springer Verlag.

Covington, M. A. (1994). Natural Languczge Processing for Prolog Programmers. Englewood Cliffs, New Jersey: Prentice Hall Inc.

Covington, M.A., and Weinrich, KB. (1991). "Unification-Based Diagnosis of Language Leamers' Syntax Errors." Literary and Linguistic Computing, 6 (3): 149-154.

Corder, S. (1967). "The Significance of Leamer's Errors." International Review of Applied Linguistics, 5: 161-170.

Designed Intelligence: A Language Teacher Mode1 153

Page 168: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Costa, E., ed. (1992). New Directions for Intelligent Trctoring Systems NATO AS1 Series F: Computer and Systems Sciences. Springer Verlag. Vol: 91.

Elsom-Cook, M. (lW3 ). "Student Modelling in Intelligent Tutoring Systems." The A r t i f ~ i a l Intelligence Reuiew, 7 (3-4): 227- 240.

Elsom-Cook, M. (1988). "Guided Discovery Tutoring and Bounded User Modelling." Artificial Intelligence and Human Learning." Self, J., ed. Bristol: J, W. Arrowsmith Ltd.: 165-178.

-

Evans, R., and Gazdar, G. (1990). m e DATR Papers, Volume 1. Brighton, School of Cognitive and Computing Sciences. The University of Sussex.

Fischer, P. M., and Mandl, H. (1988). "Improvement of the Acquisition of Knowledge by Informing Feedback." Learning Issues for Intelligent Tutoring System. Mandl, H. and Lesgold, A., eds. Springer Verlag: 187-241

Farghali, A. (198 9). "A Model for Intelligent C omputer Assisted Language Instruction (MICALI)." Cornputers and the Humanities, 23: 235-250.

Fotos, S., ed. (1996). Multimedia Language Teaching. Tokyo, Logos International,

Frasson, C., Gauthier, G-, and McCda, G.I., eds. (1992). ~ n t e l l i ~ e n t Tutoring Systems. Lecture Notes in Computer Science. Springer Verlag.

Garrett, N. (1995). 'WALL and Second Language Acquisition." Intelligent Language %tors: Theory Shaping Technology. Holland, M . V., Kaplan, J.D ., Sama, M.R., eds. Mahwah, New Jersey: Lawrence Erlbaum Associates, Inc. 344-358.

Garrett, N. (1987). "A Psycholinguistic Perspective on Grammar and CALL." Modern Media in Foreign Language Education: Theory and Implementation. Ed. Smith, W . F. L ~ ~ c o ~ ~ w o o ~ ~ K: National Textbook. 169-196.

Gazdar, G. (1982). Phrase Structure Grammar. The Nature of Syntactic Representution. Jacobson, P . and Pdum, G, eds. Dordrecht: Reidel: 13 1-186.

Designed Intelligence: A Language Teacher Mode1 154

Page 169: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Gazdar, G., and PuUum, G. (1985). "Computationally Relevant Properties of Natural Languages and their Grammars." New Generation Computing. 3: 273-306.

Gazdar, G., Klein, E., Pullum, G-, and Sag, 1. (1985). Generulked Phrase Structure Grammar. Cambridge, MA: Harvard University Press.

Greer, J. E., and McCalla, G.I., eds. (1994). Student Modelling: The Key to Individualizec? Knowledge-Based Instruction. Nato AS1 Series F: Computer and Systems Sciences. Vol: 125. Springer Verlag.

Greer, J. E., and McCalla, G.I. (1989). "A Computational Framework for Granularity and its Application to Educational Diagnosis ." Proceedings of the International Conference on Artificial Intelligence. Detroit: 477-482.

Hagen, L. K. (1994). "Unification-Based Parsing Applications for Zntelligent Foreign Language Tutoring Systems." CALICO Journal, 12 (2): 5-3 1.

Holland, M. V., Kaplan, J.D., and Sama, M.R., eds. (1995a). Intelligent Language Tutors: Theory Shaping Technology. Mahwah, New Jersey: Lawrence Erlbaum Associates, hc.

Hol la~~d , M. V., and Kaplan, J.D. (1995b). Naturd Language Processing Techniques in Cornputer-Assisted Language Learning: Status and Instructional Issues. Instructional Science, 23: 35 1-380.

HoIland, M. V. (1994). uLessons Leamed in Designing Intelligent CALL: Managing Communication across Disciplines ." Computer Assisted Lang-uage Learning, 7 (3): 227-256.

Houand, M., Maisano, R., Alderks, C., and Martin, J. (1993). "Parsers in Tutors: What Are They Good For?" CALICO Journal, 11 (1): 28-46.

Holt, P., D u ~ s , S., Jones, M., Greer, J. (1994). "The State of Student Modelling." Student Modelling: The Key to Indiuidualized Knowledge-Based Instruction. Greer, J . E., McCalla, G, eds. NATO AS1 Series. Series F: Computer and Systems Sciences. Springer Verlag. Vol: 125: 39-35.

Designed Intelligence: A Language Teacher Mode1 155

Page 170: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Holt, P., and Wood, P. (1990). "Intelligent Tutoring Systems: A Review for Beginners." Canadian Journal of Educational Communications, 19 (2): 107- 23.

Johnson, R. (1983). Parsing with Transition Networks. Parsing Natural Language. King, M. (ed). Academic Press Inc., London: 59-73.

Juozulynas, V. (1994). "Errors in the Compositions of Second-Year German Students: An Empirical Study for Paner-Based ICALI." CALICO Journal, 12 (1): 5.17.

Karttunen, L. (1984). "Features and Values." Proceedings of the International Conference on Cornputational Linguktics.

Kenning, M.-M., and Kenning, M. J. (1990). Cornputers and Language Leaming. Ellis Horwood Limited, England.

King, M., ed. (1983). Parsing Natural Language. London: Academic Press Inc.

Kunip, M., Greer, J. E., and McCda, G. (1992). The Fawlty Article Tutor." Intelligent Tutoring Systems. Frasson, C., Gauthier, G-, McCaIla G.1, eds. Lecture Notes in Computer Science. Springer Verlag: 84-91.

Kurtz, B. L., Chen, L., and Huang, X (1990). English-Chinese Language Learning System Using Adaptive Correction and Multipass Error Diagnosis." J o u m l ofArtificial Intelligence in Education, 1 (4): 51-64.

Labrie, G., and Singh, L.P.S. C19911. Tarsing, Emor Diagnostics, and Instruction in a French Tutor." CALICO Journal, 9: 9-25.

Lado, R. (1957). Linguistics across Cultures. Ann Arbor: University of Michigan Press.

LaReau, P., and Vockell, E. (1989). The Computer in the Foreign Language Curriculum. Santa Cruz, CA: Mitchell Publishing.

Desigxled Intelligence: A Language Tacher Mode1 156

Page 171: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Larson-Freeman, D., and Long, M. H. (1991). An Introduction to Second Langzmge Acquisition Research. London: Longman Group UK Limited.

Last, R. W. (1989). Mificial Intelligence Techniques in Language Learning. Ellis Horwood Limited, England.

Levin, L. S., and Evans, D. A. (1995). "ALICE-chan: A Case Study in ICALL Theory and Practice." Intelligent Language Tutors: Theory Shaping Technology. Holland, M. V., Kaplan, J.D., Sama, M.R., eds. Mahwah, New Jersey: Lawrence Erlbaum Assotiates, Inc.: 77-99.

Liou, H.-C. (1991). "Development of an English Grammar Checker: A Progress Report." CALICO Journal, 9: 27-56.

Littlewood, W. (1984). Foreign and Second Language Learning. Cambridge University Press.

Loritz, D. (1992). "Generalized Transition Network Parsing for Language Study: T h e GPARS System for English, Russian, Japanese, and Chinese." CAL,ICO Journal, 10 (1): 5-22.

Loritz, D. (1995). "GPARS: A Suite of Grammar Assessrnent Systems." Intelligent Language Tutors: Theory Shuping Technology. Holland, M. V., Kaplan, J.D., Sama, M.R., eds. Mahwah, New Jersey: Lawrence Erlbaum Associates, Inc.: 77-99.

Mandl, H., and Lesgold, A-, eds. (1988). Learning Issues for Intelligent Tutoring Systems. Springer Verlag.

Marshall, D. V. (1988). CALICBT - The Great Debate. ChartmeIl-Bratt, Brornley.

Matthews, C. (1994). "Intelligent Computer Assisted Language Learning as Cognitive Science: The Choice of Syntactic Frameworks for Language Tutoring." Journal of Artificial Intelligence in Education, 5 (4); 533-556.

Matthews, C. (1993). "Grammar Frameworks in Intelligent CALL." CALICO Journal, I l (1): 5-27.

- -

Designed Intelligence: A Language ~Acher Mode1

Page 172: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

McCalla, G. (1992). The Centrality of Student Modelling to Intelligent Tutoring." New Directions for Intelligent Trctoring Systems Costa, E., ed. NATO AS1 Senes F: Computer and Systems Sciences. Springer Verlag. Vol: 91: 107-131.

McCalla, G., and Greer, J. (1994). "Granularity-Based Reasoning and Belief Revision in Student Models." Student Modelling: The Key to Individualized Kiiowledge-Based Instruction. Greer, J. E., McCalla, G, eds. NATO AS1 . Series F. Computer and Systems Sciences. Springer Verlag. Vol: 125: 39-62.

MCC&, G., and Greer, J. (1992). "Special Issue on Student Modelling: Editors' Introduction." Journal of Artificial Intelligence in Education, 3 (4): 377-380.

McFetridge, P., and He*, T. (1995). "Grammar and Analysis for Implementing Granula* in Cornputer-Assisted Language Instiuction." Educational Multimedia and Hypennedia, 1995. Graz, Austria: Association for the Advancement incomputing in Education, 1995. pp. 460-465.

Nagata, N. (1996). "Computer vs. Workbook Instruction in Second Language Acquisition." CALICO Journal, 14 (1): 53-75.

Nagata, N. (1995). "An Effective Application of Natural Language Processing in Second Language Instruction." CALICO Journal, 13 (1): 47-67.

Nagata, N. (1991). "Intelligent Computer Feedback for Second Language -

Instxuction." The Modern Language Journal, 77: 330-8.

Nerbonne, J., Netter K., and Pollard C., eds. (1994). G e n a n in Head-Driven Phrase Structure Grammar. CSLI Lecture Notes. Stanford, Center for the Study of Language and Information.

Newmark, P. (1976). "The Importance o f Accuracy." CILT reports and Papers 13, Gennan in the United Kingdom, Problems and Prospects.

Oxford, R. L. (1995). "Linking Theories of Learning with Intelligent Cornputer-Assisted Language Learning (CALL)." Intelligent Langrcage Tutors: Theory Shuping TechnoZogy. Holland, M. V., Kaplan, J.D., Sama, M.R., eds. Mahwah, New Jersey: Lawrence Erlbaum Associates, Inc. 359-369.

Designed Intelligence: A Language Teacher Mode1

Page 173: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Pennington, M. C., ed. (1996). The Power of CALL. Houston: Athelçtan Publications.

Pereira, F.C.N., and Warren, D. H. D. (1980). "Definite Clause Grammars for Language Analysis - a Survey of the Fomalism and a Cornparison with Augmented Transition Networks." Artificial Intelligence, 13: 231-278.

Pollard, C., and Sag, 1. (1994). Head-Driuen Phrase Structure Grammr. Chicago UGversity Press.

-

Pollard, C., and Sag, 1. (1987). .hformtwn-Based Syntax and Semantics: Fundarnentals. CSLI Lecture Notes. Stanford, Center for the Study of Language and Information.

Price, R. V. (1991). Cornputer-Aided Instruction: A Guide for Authors. Pa&c Grove, CA: BrooWCole Publishing Co.

Richards, J. (1971). "Error Analysis and Second Language Strategies." Languuge Sciences, 17: 12-22.

Rogers, M. (1984). UOn Mayor Types of Written Error in Advanced Students of German-" International Review of Applied Linguistics in Languuge Teaching, 22 (1): 1-39.

Rumelhart, D., and Norman, DA. (1975). Explorations in Cognition. San Francisco: Freeman.

Sanders, R. (1991). YError Analysis in Purely Syntactic Parsing of Free Input: The Example of German." CALICO Jourrurly 9 (1): 72- 89.

Sanders, A. F., and Sanders R. H. (1989). &Syntactic Parsing: A Survey." Cornputers and the Humanities, 23: 13-30.

Schuster, E. (1986). The Role of Native Grammars in Correcting Enors in Second Language Learning." Cornputafional Intelligence, 2 (2): 93-98.

Schwind, C. B. (1995). "Error Analysis and Explanation in Knowledge Based Language Tutoring." Cornputer Assisted Languuge Learning, 8 (4): 295-325.

Designed Intelligence: A Language Teacher Mode1 159

Page 174: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Schwind, C. B. (1990a). "An Intelligent Language Tutoring System." international Journal of Man-Mcchine Studies, 33: 557-579.

Schwind, C. (1990b). Teature Grammars for Semantic Analysis ." Computational Intelligence, 6: 172-178.

Self, J., ed. (1988). Artificiul Intelligence and Human Learning- London, New York: Chapman and Hall Ltd.

SelinJcer, -L. (1972). "Interlanguage." International Eeview of Applied Lingilistics, 10: 2091231.

Sevin, D., Sevin, I., and Bean K.T. (1991). Wie gehfs. Holt, Rinehart and Winston, Inc. The Dryden Press. Saunders College Publishing.

Shieber, S. M. (1986). An Introduction to Unîf~aion-Based Approaches to Grammar. CSLI Lecture Notes Number 4. Stanford, Center for the Study of Language and Information.

Sleeman, D., and Brown, J. S. eds. (1982). Intelligent Btoring Systems. Academic Press,

Smith, W. F., ed. (1987). Modern Media in Foreign Language Education: m o r y and Impkmentation. Chicago: National Textbook Co.

Swartz, M. L., and Yazdani, M., eds. (1992). Intelligent Tutoring Systems for Foreign Language Leanzing. The Bridge tu Intenational Communication. Nato AS1 Series F: Computer and Systems Sciences. Vol: 80. Springer Verlag.

Van der Linden, E. (1993). "Do es Feedback Enhance Cornputer-Assisted Language Learning." Computers & Education, 21 (1-2): 61-65.

Van Noord, G. (1997). "An Efficient Implementation of the Head-Corner Parser." Computational Linguistics, 23 (3): 425-456.

Venetzky, R., and Osin, L. (1991). The Intelligent Design of Computer-Assisted Instruction. New York: Longman Publishing Group.

Wang, Y., and Garigliano, R. (1995). "Empirical Studies and Intelligent Language Tutoring." Instnrctional Science, 23: 47-64.

Designed Intelligence: A Language Teacher Mode1 160

Page 175: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

Wang, Y., and Garigliano, R. (1992). "An Intelligent Language Tutoring System for Handling Errors Caused by Tramfer." Intelligent Tutoring Systems. Frasson, C., Gauthier, G., McCalla G.1, eds. Lecture Notes in Cornputer Science. Springer Verlag: 395-404.

Warschauer, M. (1996). "Computer-Assisted Language Learning: An Introduction." In Fotos, S. (ed). Multimedia Langxage Teaching. Tokyo: Logos International: 3-20.

Weinreich;-U. - - . (1953). L a m a g e in Contact. Linguistics Circle of New York.

~eischedel , R.M., and Sondheimer, N.J. (1983). "Meta-Rules as a Basis for Processing Ill-Formed Input." American Journal of Computational Linguistics, Special Issue on I l l -Fomd Input, 9 (3-4): 161-177.

Weischedel, R.M., Voge, W.M., and James, M. (1978). "An Artincial Intelligence Approach to Language Instruction." Artificial Intelligence, 10: 225-240.

Wenger, E. (1987). Artifziul Intelligence and Tutoring Systems: Computational and Cognitive Approaches to the Communication of Rnowledge. Los Altos, CA: Morgan K a h a n n Publishers, Inc.

Woods, W. (1970). "Cascaded ATN Grammars." American Journal of Computational Linguistics, 6 (1): 1-12.

Yazdani, M., and Uren, J. (1988). "Generalising Laquage-Tutoring Systems: A French/Spanis h Case Study, Using LINGER." Instructional Science, 17: 179-188.

Designeci Intelligence: A Language Teacher Mode1 161

Page 176: INFORMATION TO USERS - Library and Archives Canada · of the possible parses by taking into account the likelihood of the error. ... 1.1.2.1 Anticipating Ill-Formed Student Input

IMAGE EVALUATION TEST TARGET (QA-3)

APPLlE D IMAGE. lnc - = 1653 East Main Street - --2 Rochester. NY 14609 USA -- -- - - Phone: 71 6/482-O3OO -- -- - - Fa: 71 6/288-5989

0 1993. Appiied Image. Inc.. All Rights Reserved