Translation quality assessment redefined

63
TRANSLATION QUALITY ASSESSMENT REDEFINED from TQI to competences and suitability Demid Tishin All Correct Language Solutions www.allcorrect.ru

Transcript of Translation quality assessment redefined

Page 1: Translation quality assessment redefined

TRANSLATION QUALITY ASSESSMENT REDEFINEDfrom TQI to competences and suitability

Demid Tishin

All Correct Language Solutions

www.allcorrect.ru

Page 2: Translation quality assessment redefined

What are they thinking about when they look at the target text?

2

Page 3: Translation quality assessment redefined

Client:

Will it blend?* Let’s find a flaw…

33

*Just a joke. “Will it do”, I mean

Page 4: Translation quality assessment redefined

Quality manager:

Will it blend? I wish the client said

OK…

4

Page 5: Translation quality assessment redefined

HR / Vendor Manager:

What kind of work can I trust to this provider? What can I

not?How quickly can we train him?

5

Page 6: Translation quality assessment redefined

Project Manager:

Return for improvement or correct by other

resources?

6

Page 7: Translation quality assessment redefined

To answer these questions the target text needs

assessment

7

Page 8: Translation quality assessment redefined

TRANSLATION ASSESSMENT: THE ARMORY

What assessment techniques do you know?

8

Page 9: Translation quality assessment redefined

TRANSLATION ASSESSMENT: THE ARMORY

Subjective assessment (“good / bad”)

Comparing with the sourceaccording to a parameter checklist

Automated comparison with a reference translation (BLEU etc.)

Weighing errors and calculating TQI 9

Page 10: Translation quality assessment redefined

SUBJECTIVE ASSESSMENT (“GOOD / BAD”)

Pro’s Con’s

Speed Results not repeatable

Results not reproducibleDifficult for client and service provider to arrive at the same opinion

Impossible to give detailed reasons

Tells nothing of provider’s abilities

10

Page 11: Translation quality assessment redefined

COMPARING WITH THE SOURCEACCORDING TO A PARAMETER CHECKLIST

Pro's Con's

Some reasoning for assessment results

Results not reproducibleDifficult for client and service provider to arrive at the same opinion

Results not repeatable

Tells nothing of provider’s abilities 11

Page 12: Translation quality assessment redefined

AUTOMATED COMPARISON WITH A REFERENCE TRANSLATION

The more word sequences correlate between the target and the reference, the better the translation

BLEU (BiLingual Evaluation Understudy), ROUGE, NIST, METEOR etc.

An overview of BLEU: Tomedes Blog http://blog.tomedes.com/measuring-machine-translation-quality/

12

Page 13: Translation quality assessment redefined

AUTOMATED COMPARISON WITH A REFERENCE TRANSLATION

13

Page 14: Translation quality assessment redefined

AUTOMATED COMPARISON WITH A REFERENCE TRANSLATION

Pro's Con's

Speed Does not account for individual style

Limited scope (today limited to MT output)

Does not correlate to human assessment

A number of reference translations must be prepared before assessment (justified for batch assessment of different translations of the same source sample)

Tells nothing of provider’s abilities

How should acceptability threshold be defined? 14

Page 15: Translation quality assessment redefined

WEIGHING ERRORS AND CALCULATING TQI

Who? Lionbridge, Aliquantum, Logrus, All Correct Language Solutions … and many others

Publicly available techniques:- SAE J2450- ATA Framework for Standard Error Marking- LISA QA Model 3.1

An overview of translation quality index techniques and guidelines to create your own:http://www.aliquantum.biz/downloads.htm 15

Page 16: Translation quality assessment redefined

WEIGHING ERRORS AND CALCULATING TQI

What components you will need: Error classifier Error weighing guidelines Translation assessment guidelines, which yield

repeatable and reproducible results Expert (competent and unambiguous) Assessment results form

16

Page 17: Translation quality assessment redefined

WEIGHING ERRORS AND CALCULATING TQI

17

Page 18: Translation quality assessment redefined

WEIGHING ERRORS AND CALCULATING TQI

18

Page 19: Translation quality assessment redefined

WEIGHING ERRORS AND CALCULATING TQI

19

Page 20: Translation quality assessment redefined

WEIGHING ERRORS AND CALCULATING TQI

TQI (Translation Quality Index) is the usual practical result of translation quality measurement

ATA Framework: TQI = EP * (250 / W) - BP

SAE J2450: TQI = EP / W

LISA QA Model: TQI = (1 - EP / W) * 100

where EP = total Error Points

W = number of words in sample

BP = Bonus Points for outstanding translation passages (ATA)

max. 3 points20

Page 21: Translation quality assessment redefined

WEIGHING ERRORS AND CALCULATING TQI

Pro's

Results highly reproducible (SAE J2450)

Results highly repeatable (SAE J2450)

Detailed error classifier with explanations and examples (LISA QA Model)

Easy to use for quality feedback to providers

Convenient to grade providers according to their TQI for a specific project

TQI is a simple numeric index, which you can account in a database and use in your balanced scorecard, KPI’s etc. 21

Page 22: Translation quality assessment redefined

WEIGHING ERRORS AND CALCULATING TQI

Con's

Limited scope (SAE J2450)

Low reproducibility of results (ATA Framework)

A threshold of acceptable TQI is required (e.g. 94,5 etc.), while clients do not tolerate any explicitly stated imperfection

Assessment is time-consuming (5-20 minutes per sample provided that the expert has carefully studied the source.

Subjective or underdeveloped error weight assignment – a try at forecasting error consequences (LISA QA Model)

Tells very little of provider’s abilities 22

Page 23: Translation quality assessment redefined

WEIGHING ERRORS AND CALCULATING TQI

Con's

Underdeveloped Translation assessment guidelines, including but not limited to: - requirements to translation sample (size, presence of terminology etc.)- how to evaluate repeated typical (pattern) errors?- how to assess flaws in the target, which root in obvious flaws in the source?- how to evade several minor errors resulting in the same score as one major error?- how to handle obviously accidental errors that change the factual meaning? 23

Page 24: Translation quality assessment redefined

WEIGHING ERRORS AND CALCULATING TQI

Con's

TQI is valid only for:- a specific subject field (e.g. gas turbines, food production etc.)- a specific text type (Legal, Technical and Research, or Advertising and Journalism)

A slight change in any of the above (subject, text type) means that one cannot forecast the provider’s TQI based on former evaluations a new (tailored) assessment is required ungrounded expenses 24

Page 25: Translation quality assessment redefined

None of the translation assessment methods answers the questions:

Will it blend?What kind of work can I trust

to this provider? What can I not?How quickly can we train him?Return for improvement or correct by other

resources?

TRANSLATION ASSESSMENT: THE ARMORY

25

Page 26: Translation quality assessment redefined

Translation assessment techniques

need improvement!

TRANSLATION ASSESSMENT: THE ARMORY

26

Page 27: Translation quality assessment redefined

Split all errors into 2 major groups:Factual = error in designation of objects and

phenomena, their logical relations, and degree of event probability / necessity

Connotative = errors in conveying emotional and stylistic information, non-compliance with rules, standards, checklists and guidelines etc.

IMPROVEMENT 1: TWO ERROR DIMENSIONS

27

Page 28: Translation quality assessment redefined

IMPROVEMENT 1: TWO ERROR DIMENSIONS

28

That’s a restaurant

That’s a damn fctory

That’s a factory

(factual error)

(2 connotative errors, though no factual errors)

(source)

Page 29: Translation quality assessment redefined

Each text element (word, phrase, sentence etc.) can contain:

1 connotative error

or1 factual error

or1 connotative and 1 factual error

simultaneously

IMPROVEMENT 1: TWO ERROR DIMENSIONS

29

Page 30: Translation quality assessment redefined

An accidental error (e.g. an obvious typo), which leads to obscuring factual info, counts as two (e.g. language and factual).

You can at once give specific instructions to the provider (e.g. be more careful) and consider client’s interest (e.g. absence of factual distortions whatever the reason)

“To kill Edward fear not, good it is” / “To kill Edward fear, not good it is” (Isabella of France): an error in the comma critical factual distortion

IMPROVEMENT 1: TWO ERROR DIMENSIONS

30

Page 31: Translation quality assessment redefined

Map each error in the classifier to the competences that are required to avoid it

IMPROVEMENT 2: COMPETENCES

31

Page 32: Translation quality assessment redefined

Competence types:Competences of acquisitionCompetences of productionAuxiliary (general) competences

IMPROVEMENT 2: COMPETENCES

32

Page 33: Translation quality assessment redefined

Competence levels:Unsatisfactory = provider cannot do the

corresponding workBasic = can workAdvanced = can revise and correct work

of others or train others

IMPROVEMENT 2: COMPETENCES

33

Page 34: Translation quality assessment redefined

Competences of acquisitionSource language rulesSource Literary Source CulturalSubject matter

IMPROVEMENT 2: COMPETENCES

34

Page 35: Translation quality assessment redefined

Competences of production:Target language rulesTarget literaryTarget culturalTarget mode of expression (= register,

functional style)

IMPROVEMENT 2: COMPETENCES

35

Page 36: Translation quality assessment redefined

Auxiliary (general) competences:ResearchTechnicalGeneral Carefulness, Responsibility and

Self-organisationCommunication (relevant for translation

as a service, not the product)Information security (relevant for

translation as a service, not the product)

IMPROVEMENT 2: COMPETENCES

36

Page 37: Translation quality assessment redefined

IMPROVEMENT 2: COMPETENCES

37

Page 38: Translation quality assessment redefined

IMPROVEMENT 2: COMPETENCES

38

Page 39: Translation quality assessment redefined

IMPROVEMENT 2: COMPETENCES

Client can formulate precise and objective requirements to the provider

Assessment immediately shows which competences stand to the required level and which don’t

39

Page 40: Translation quality assessment redefined

IMPROVEMENT 3: WORKFLOW ROLES

Map each workflow role (e.g. translate, compile project glossary, revise language etc.) to a number of required competences

40

Page 41: Translation quality assessment redefined

IMPROVEMENT 3: WORKFLOW ROLES

Example of a competence set: Self-organisation = basic Subject matter = basic Source language rules = basic Target language rules = basic

Role: Can translate

41

Page 42: Translation quality assessment redefined

IMPROVEMENT 3: WORKFLOW ROLES

42

Page 43: Translation quality assessment redefined

IMPROVEMENT 3: WORKFLOW ROLES

Vendor Manager / Project Manager quickly assigns workflow roles and schedules the project

saves time

43

Page 44: Translation quality assessment redefined

IMPROVEMENT 4: ERROR ALLOWABILITY

In each case client indicates which error types are allowed and which are not. The expert puts down the client requirements in a list

One “not allowed” error in the sample

Text fails (client perspective)

44

Page 45: Translation quality assessment redefined

IMPROVEMENT 4: ERROR ALLOWABILITY

45

Page 46: Translation quality assessment redefined

IMPROVEMENT 4: ERROR ALLOWABILITY

Assessment stands to the real client needs (“pass / fail”)

46

Page 47: Translation quality assessment redefined

IMPROVEMENT 5: PROVIDER TRAINABILITY

Single out 2 major error groups: Correcting the error requires minimum training /

instructions; the provider can find and correct all errors of the type in his work himself

Correcting the error requires prolonged training; the provider cannot find all his errors in the text

47

Page 48: Translation quality assessment redefined

IMPROVEMENT 5: PROVIDER TRAINABILITY

Errors that require minimum training: the original order of text sections is broken broken cross-references text omissions numbers / dates do not correspond to the source glossary / style guide violated non-compliance with reference sources inconsistent terminology non-compliance with regional formatting standards broken tags, line length obvious language errors etc.

48

Page 49: Translation quality assessment redefined

IMPROVEMENT 5: PROVIDER TRAINABILITY

Errors that require prolonged training:understanding the sourceconfusion of special notions

(subject competence) stylistic devices and expressive means

(literary competence)cultural phenomenaetc.

49

Page 50: Translation quality assessment redefined

IMPROVEMENT 5: PROVIDER TRAINABILITY

50

Page 51: Translation quality assessment redefined

IMPROVEMENT 5: PROVIDER TRAINABILITY

What is the percentage of errors requiring minimum training?

PM can instantly take a decision – return the product for further improvement or correct with other resources

saves time51

Page 52: Translation quality assessment redefined

IMPROVEMENT 5: PROVIDER TRAINABILITY

If all errors influencing the competence are easy to correct, the competence is assessed in two ways (at once):

Current state (“as is”)Potential state (after a short training)

52

Page 53: Translation quality assessment redefined

NOTES

Provider has to work in normal conditions (enough time, work instructions)

The sample should be restricted to one main subject field according to the subject classifier

The source text should be meaningful and coherent

It is important to differentiate between errors and preferential choices

53

Page 54: Translation quality assessment redefined

NOTES

To assess a sample the expert has to possess all competences on "advanced" level.

As it is difficult to find such experts in reality, several people can be assigned to assess one sample (e.g. one assesses terminology, another assesses all the other aspects)

Quality predictions for rush jobs cannot be based on normal competence assessment (as the rush quality output is normally lower)

54

Page 55: Translation quality assessment redefined

EXAMPLE

55

Page 56: Translation quality assessment redefined

EXAMPLE

56

Page 57: Translation quality assessment redefined

EXAMPLE

57

Page 58: Translation quality assessment redefined

EXAMPLE

58

Page 59: Translation quality assessment redefined

CONCLUSION

The new assessment model replies to all the questions:

Will it blend? – pass / fail What kind of work can I trust to this provider?

What can I not?– competences and workflow roles How quickly can I train the provider? – potential

competences Return for improvement or correct by other

resources? – percentage of errors requiring minimum training

59

Page 60: Translation quality assessment redefined

BENEFITS Provider and client speak the same “language”

(error types and competences) less debates Saves time when testing providers Simplifies planning of minimum and sufficient

workflow, optimizes resources Allows to avoid extra text processing stages

when not necessary (extra stages avoided) better turnaround (extra stages avoided) more flexible budgets

higher rates provider loyalty and good image for the company

Detailed feedback and training provider loyalty

60

Page 61: Translation quality assessment redefined

THE FUTURE OF THE TECHNIQUE

Adjustment and testingDedicated software toolIntegration with QA tools

61

Page 62: Translation quality assessment redefined

62

QUALITY MANAGEMENT PROCESS

Page 63: Translation quality assessment redefined

Thank you!

Questions?

[email protected] 63