Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents...

47
9KK=KKE=FL 9KK=KKE=FL e\ e\ Yecf[j[dY[ Yecf[j[dY[

Transcript of Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents...

Page 1: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

99KKKK==KKKKEE==FFLL99KKKK==KKKKEE==FFLLee\\ ee\\

YYeeccff[[jj[[ddYY[[YYeeccff[[jj[[ddYY[[

Page 2: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy
Page 3: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

C o n t e n t sPPaaggee NNoo.. 11 -- 22

Introduction 3

Overview 5

Competence 6

Systems approach 9

Taxonomy of educational objectives 11

Guidelines for stating specific learning outcomes 15

Table of specifications 17

Principles of assessment 20

Characteristics of a good examination 21

Types of validity 23

Essay questions: characteristics 24

Essay questions 25

Short essay questions 26

Scoring strategies 29

Analytical vs holistic rubrics 33

Options for selecting rubrics 35

Analytical rubric: example 40

Blank rubric template for practice 43

1Assessment of Competence

DAY

ONE

Page 4: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

ASSESSMENT OF COMPETENCE

S C H E D U L E

DAY ONE

Inaugural

Problems with assessment in Pakistan

What do we need to assess

Why do we need to assess

How can we improve assessment

Making a TOS

Principles of Assessment

Characteristics of a good exam

Validity and its types

Essay questions

Characteristics of short essays

How to write short essays

Critique

Scoring strategies

Analytical vs. Holistic rubrics

Making rating scale for short essays

Critique of rating scale

2Assessment of Competence

DAY

ONE

Page 5: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

INTRODUCTIONThe College of Physicians & Surgeons Pakistan (CPSP) was established in1962 for the purpose of training local doctors in higher medical education.Its founder’s spelt out the following objectives for setting up thisinstitution:

By the end of the workshop, you will be able to:l to promote specialist practice in medicine, surgery, gynecology and

obstetrics and such other specialties;l to arrange postgraduate medical training;l to hold and conduct examinations for grant of diplomas and admission

to the Fellowship;l to provide for research.

Today, CPSP is recognised as one of South Asia’s major postgraduatemedical education centres with an international reputation in education,training and research. It has strong linkages with leading sister institutionsabroad.

Traditionally postgraduate training in Medicine in Pakistan used to followthe pattern as in the United Kingdom (UK). However, with the passage oftime the College has evolved its own system of training and evaluation toproduce a large pool of internationally recognised specialists who areinstrumental in raising the standard of health care in the country. TheCPSP has produced more than 10,000 specialists till date.

Being a premier institution, CPSP is acutely aware of the quality of itstraining and assessment. It has consistently been taking steps forimproving all aspects of the degree that it awards. One of the key steps inimproving the quality of its training is the training of its existing and to-be-supervisors in prime facets of imparting training. This includes planning atraining event, implementing it and then evaluating it.

The four newly designed workshops are focused around these threeaspects. Here is a brief overview of each of these:

Educational Planning and evaluation:This is geared towards how to plan an academic and structured trainingprogram. Devising a structured training program, or for that matter anyacademic event, a planned event and goes beyond making a time table orschedule. It requires being aware of the needs of the learner and thesociety, being aware of the steps in planning and the basic steps inassessing the competence of the learner.

3Assessment of Competence

DAY

ONE

Page 6: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

This workshop is the first in series of workshops that a clinician has toattend. It is hoped that in this first exposure. The participants will be ableto devise and plan effective academic events.

Assessment of competence:Assessment is a tricky field to play on. It requires careful planning. If onefollows the procedures outlined in Educational Planning & Evaluation andcarries them out, the following steps for assessment of competence willbecome easy for the clinician and supervisor. One can not but agree thatin most cases, assessment does drive learning. Keeping this perception inmind, a supervisor has to be adept at planning and selecting tools forgauging competence. This workshop is directly linked to the previous oneand is intended to equip the participants with important skills of writingshort essay questions, one best and extended matching. It will also helppeople develop and run TOACS to assess psychomotor skills. Assessmentof Competence is important also since it endeavors to explain the qualityassurance procedures that any academic forum should follow in order toproduce reliable and valid decisions about competence or lack thereof.

Supervisory Skills:Having learnt about planning academic experiences it becomes imperativefor the supervisor to learn how to implement the whole plan. This workshopdiscusses practical measures that a clinician has to take when he / sheassumes the responsibility of being a trainer of fellow ship programs. Ithelps the participant gain experience of dealing with practical issues likeconflict management, training juniors in counseling, implementing atraining program.

Research Methodology and Medical Writing:An important aspect of supervision is organizing the conduct of researchof trainees for their dissertation and articles which are a pre- requisite forsitting in fellowship exams.In this workshop, supervisors get to learn the format and standards CPSPexpects from researchers.

All workshops lay great emphasis on hands on training and frequentinteraction between participants and facilitators. Participants of the workshops need to follow the sequence as:But there is a sequence that has to be followed in the remainingworkshops:1st Educational Planning and Evaluation, then Assessment of Competenceand finally Supervisory Skills. Research Methodology and Medical Writingmay be attended at any step.

4Assessment of Competence

DAY

ONE

Page 7: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

OVERVIEWThe Department of Medical Education, College of Physicians & SurgeonsPakistan is interested in raising the standard of medical education.Assessment of student achievement is one of the most importantresponsibilities of teachers. Both written tests and per formanceassessments serve many functions, but at the same time poorly conceivedand executed tests may do more disservice to the learners. Therefore it isimportant that assessment instruments be well thought out and wellmade. It is also important for examiners to know what to do with theresults. It is not sufficient to simply paste results on the notice board withpercentages since it does not serve much purpose.

The present workshop is a continuation of the same effort and is beingundertaken to introduce to a larger number of faculty members of variousmedical colleges, issues related to two very important techniques ofassessment i.e. short essays and MCQs. Among MCQs, we lay stress onOne best and Extended Matching since internationally and nationally theyare proving to be better than true false. This is because the former two arecapable of gauging higher cognitive thinking much better than the latter.Also chances of guessing and cues are reduced (but not eliminated). Alsohands-on experience will be provided in developing and using tools forassessing psychomotor sheets and affects.

This workshop is based on mutual group activity and it is important for itssmooth running that everyone participate fully. It is for this reason that youare requested to attend all the sessions and avoid any other commitmentduring these hours.

Note: No telephone calls or messages will be delivered during thesessions. Please keep your mobiles and pagers switched off while thesessions are in progress.

OBJECTIVES

By the end of the workshop, participants will be able to:

By the end of the workshop, you will be able to:l Write objectives sutable for their training programl Develop a table of specifications l Develop an assessment plan based on characteristics of a good

examinationl Develop analytical rating scales for assessing cognition, procedural

skills and affectsl Write Short Essays questions, One best and Extended Matching MCQsl Organize and conduct examinations for the assessment of Clinical skillsl Interpret analysis of results based on reliability, dif ficulty and

discrimination indicesl Discuss the uses of an exam result

5Assessment of Competence

DAY

ONE

Page 8: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

COMPETENCE

Is defined as an individual’s demonstrated knowledge, skills and abilities(KSAs) performed to a specific standard. Competencies are observable,behavioral acts that require a combination of KSA’s to execute. It is

necessary to have them documented for effective performance of a job ortask.

CORE COMPETENCIES:Certain competencies are essential to all members regardless of thediscipline in which they are. For example, history taking, patientexamination, communication with patient, requesting relevantinvestigations, interpretation of various lab data etc. these are factorswhich all disciplines will lay great stress on. These are the skills withoutwhich no person can be certified as a safe and ‘competent practitioner.

SPECIFIC COMPETENCIES:Are those skills, knowledge and abilities that are going to be usedspecifically in a particular discipline. For example the competenciesrequired in a histopathology department will be slightly different from thatin a neonatal care unit. Similarly, there will be a substantial difference inthe competencies required in a Psychiatry unit as compared to anOphthalmology unit.

SOFT-SKILL COMPETENCIES:Are competencies that are required for acquiring skill in auxiliary areas like l Computer skills,l Management skills, l Knowledge of research skills, l Time management l Conflict management l Negotiation skills etc.

Such competencies considered essential for the smooth running of thetraining unit many clinicians consider.

6Assessment of Competence

DAY

ONE

Page 9: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

CLINICAL COMPETENCIESApplication of Knowledge= The application of clinical and basic science knowledge in order to

diagnose and manage a patient's medical problems.

History Taking= The technical skills involved in interviewing a patient. The skill of asking

questions relevant to the patient and building on to the informationbeing provided. A constant analysis and expertise a required by anadept clinician.

Physical Exam Technique (Data Collection)= Conducting a satisfactory physical examination. Includes knowing all

the physical examination maneuvers,

= the dexterity of conducting an examination efficiently with minimalduress to the patient is an important component of competence.

Clinical reasoning = Recognition of significant findings established through questions

(history), inspection, palpation, Auscultation, and percussion.

Interpretation of clinical data= The ability to integrate information about the patient and arrive at a

reasonable differential diagnosis.

Plan of investigation = Systematic choice of laboratory tests and other diagnostic

investigations.

Diagnostic procedures= Ability to perform diagnostic procedure.

Patient Problem Identification= Integrating the information in a most plausible problem.

Procedural skills= The ability to perform a technical procedure with causing none or

minimal discomfort to the patient

Patient Management= The ability to plan and carry out an appropriate approach to manage the

patient's health related problem.

7Assessment of Competence

DAY

ONE

Page 10: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

Communication= The ability to communicate clearly and effectively with patients and

colleagues in both written and spoken forms.

Counseling= The ability to assist the patient arrive at a decision on his / her own

effectively regarding the nature of the medical problem; the plan fordealing with the problem and the patient's role in management.

Professional Behavior= The ability to interact and work effectively with patients and other health

professionals. This includes key personal traits as organization,dependability, and ethical conduct.

8Assessment of Competence

DAY

ONE

Page 11: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

SYSTEMS APPROACH

Systems approach to education is a conceptual framework,

l which explains what the various components of an effective educational plan are,

l in what sequence they appear and l how they are inter-related.

It aids us in tackling problems that we experience in making our studentslearn better and hence, be more competent.

OBJECTIVES:

Objectives are statements of what the educational process hopes to

achieve. For medical education in Pakistan, they are primarily laid down by

the PMDC, which draws upon the services of health officials and experts

in different fields in designing them. The needs of the nation, community,

profession, students etc. determine the objectives.

Currently, the educational worlds is focusing on outcomes of an

educational program. Hence, the term ‘objective’ has been replaced by

‘outcome’. So, through out this text we will refer to objectives as

outcomes. Consequently, the first major step in the systems apporach is

the writing of learning outcomes. Since these have to be written in specific

and measureable terms, they are called Specific Learning Outcomes or

SLOs. SLOs inform the readers what the learner will have accomplished

after going through the course.

INSTRUCTIONAL STRATEGY:This is the outline or plan of how the objectives should be met. In practical

terms, this is the timetable and schedule which each institution and

department prepares in order to achieve the objectives. The strategy is

partly dependent upon the resources although it is possible to maximize

the available resources by fully using the principles of learning.

EVALUATION:Evaluation, is a general term, which incorporates a number of modalities.

For example, evaluation could be to find out the quality of a teaching

program, the teaching abilities of a faculty member, the strength of a

curriculum etc.

9Assessment of Competence

DAY

ONE

Page 12: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

Assessment, on the other hand, is used specifically to mean evaluation ofstudents' knowledge, psychomotor competence and/or attitudes. Withinthe same context, assessment is the process of determining whether theobjectives have been achieved or not. A properly designed assessment isthus determined by including a sufficient number of SLOs in the test orexam; this is called sufficient sampling. Student assessment is better ifthe principles of evaluation are utilized in designing it.

The systems approach can be used in our lives also, i.e. first we make aplan (objectives) to do some work and then we perform it and then weevaluate what we did by thinking about how our work went. So usually we‘evaluate’ our work mentally by matching what we did with what we hadplanned to do. So the evaluation is always according to the initial plan. Ifthings do not go as per our plan (objectives) we review where we wentwrong. This review is not just of how we did our work (instructional strategy)but also of whether we evaluated our plan correctly or not and whether ourinitial plan was correct or not. So, review is ALWAYS of each and everyaspect of the systems approach.

REVIEW:It is imperative that once the whole circle has been followed, experts in thesubject hold a meeting in which they review the whole plan. They need to:

1. Go over the results of the test that they have administered and askthemselves: does the test have indicators of high quality? If not, thenthe items in the test have to be reviewed and

2. Matched with the objectives (do the SLOs need to be changed also?). 3. Some changes in the instructional strategy may then be necessary

and/ or4. Changes in the test that was administered.

10Assessment of Competence

DAY

ONE

Page 13: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

TAXONOMY OF EDUCATIONALOBJECTIVES

HISTORICAL BACKGROUNDEducational development has been proceeding since the 19th century. The

beginning of the 20th century saw a rapid growth in work done in the field

of education. Educational psychologists evolved their ideas on the basis of

much research that kept pouring in from all over the world. Hence, a

number of educational philosophies evolved with time. So, we had the

‘Traditional philosophy’ of education (1897); then came the ‘Experiential

philosophy’ (1931); ‘Structure of the Disciplines’ (1953) became very

popular after world war II; ‘ Behavioral philosophy’ gained prominence

during the late 1950’s; during the late 1960’s the ‘Cognitive philosophy’

became very much in vogue since, by then, scientists had learned enough

about the functioning of the mind to be able to apply the information to

learning and education.

It was actually during the mid 50’s that it was thought of classifying

knowledge and information. Bloom, in 1956, was amongst the pioneers in

presenting this classification to the world. We have tried to simplify the

extensive taxonomy or classification presented by Bloom. Following is its

description.

TAXONOMY OF EDUCATIONAL OBJECTIVESThe taxonomy of educational objectives is intended to provide for

classification of the goals of our educational system. Although the specific

learning outcomes resulting from a course of study may run into hundreds,

most of them can be classified under a small number of headings.

Any such classification serves several useful purposes. It:

l indicates the type of learning outcomes that should be considered.

l provides a framework for classifying those outcomes.

l directs attention towards changes in student performance in a variety of

areas.

l forms the blueprint for forming your assessment tool.

l provides common grounds and makes comparison of various teaching

programs easier.

11Assessment of Competence

DAY

ONE

Page 14: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

The system proposed by Bloom first divides objectives into the followingthree major areas:

Cognitive domain: Knowledge outcomes, intellectual abilities and thinkingskills.

Psychomotor domain: Perceptual and motor skills.

Affective domain: Attitudes, interests, appreciation, and modes of

adjustment.

Each of the three domains is subdivided into a number of sub-categories.

For the sake of convenience and based on literature from National Board

of Medical Education USA, the original categories and have been simplified

grouped into three sub catagries;

Following are the levels of Cognitive domain:

C1: involves recall of facts, principles, processes, patterns and methods

necessary for efficient performance of a professional task.

C2: involves deriving meaning from a single source of information.

Laboratory data, X-rays, ECGs or graphs are interpreted to determine

whether they are normal or abnormal in a given situation. Interpretation is

not possible without the knowledge of relevant facts and thus questions

aimed at testing this level automatically test C1 as well.

C3: start with gathering data from all the available sources. This data is

then interpreted, and highest synthesized and then analyzed to reach a

conclusion encompasses making diagnoses, management play, clinical

decision-making, clinical reasoning. This, at best, includes finding

solutions for a problem arising from new situations with very little or no

precedence to serve as a guide but by recalling facts and analyzing the

infomation at hand. Problem solving and clinical management. This is the

best and since it incorporates the previous two levels automatically.

12Assessment of Competence

DAY

ONE

Cognition 1 (C1): Recognition and recall: entails recall of facts and figures

Cognition 2 (C2): Interpretation: involves recall, synthesis and desisions about a single piece of information

Cognition 3 (C3): Problem Solving: consists of recall, synthesis, analysis and judgement

Page 15: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

Following are the levels of psychomotor domain:

P1) IMITATION:

This is the first level of competence of performing any skill. During this

stage the person has to devote all his attention to the work at hand and is

unable to perform complex tasks unless they are systematically broken in

to smaller tasks. He needs constant supervision and instant feedback

about his actions and mistakes, which are frequent.

P2) CONTROL:

The performer is slightly more adept at the skill. He still needs to pay attention

but the level of attention required is not as great as before. He is able to

perform complex tasks in bigger chunks and relatively more speed. At this

stage he still needs supervision but to a much less degree. The number of

mistakes that he makes has reduced. He is able to judge when he makes a

mistake but has difficulty in making correction. He also has a lot of problems

in transferring his recently acquired skill to new and unique situations.

P3) AUTOMATION:

The performer is fully skilled at the work and needs no supervision. He is

able to focus his attention on things other than the task while performing.

The frequency of his mistakes has reduced remarkably. He is able to self

analyze the mistakes he makes and correct them himself. He is also able to

use his skill in new situations with relatively more ease. He is now able to

reflect on what he has done and how he can improve upon it.

13Assessment of Competence

DAY

ONE

Psychomotor 1 (P1): Imitation: Only small and simple tasks possible, one at a time, many mistakes, high tress level, maximumattention and heavy supervision required

Psychomotor 2 (P2): Control: Slightly complex tasks, fewer mistakes, moderate or low stress/and less supervision required, inability to deal with unique situations

Psychomotor 3 (P3): Automation: Complex multi-tasking, very few mistakes, self analysis and correction, command during unique situations, no supervision required

Page 16: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

Following are the levels of Affective domain:

A1) RECEPTIVITY: This stage implies sensitivity to existence of a

certain phenomenon and includes willingness to receive. Example:

Noticing the anxiety of a patient awaiting the result of a laboratory test for

a disease that can have serious consequences.

A2) RESPONSE: This implies sufficient interest in the phenomenon

noticed to do something about it. Example: in the case described above in

the previous example, the response would be to say a few reassuring

words to that patient so that he does not feel alone.

A3) INTERNALIZATION: This level of affective domain implies that your

perception of a phenomenon has found a place in your scale of values and

has affected you long enough for you to adapt yourself to the value system

of the other person. This enables you to adapt your attitude to the other

person as if you were experiencing the same phenomenon yourself.

Example: On the death of a child, your attitude to the members of his

family will show them that you care about their grief and are ready to help

them get over it. This does not mean that you have internalized their grief

but that you have internalized the attitude that enables you to offer them

genuine, effective help.

14Assessment of Competence

DAY

ONE

Affective 1 (A1): Receptivity: Sensitivity to a phenomenon and willingness to receive

Affective 2 (A2): Response: Action in response to a phenomenon due to inherent interest

Affective 3 (A3): Internalization: Inclusion of the perception within the personal value system

Page 17: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

GUIDELINES FOR WRITING SPECIFICLEARNING OUTCOMES

Objectives are statements of intentions they have to be thought of and

documented in the initial phase of any educational activity (as per the

systems approach).

The whole of ‘Behavioral philosophy’ of learning is based on stating

intentions of how or what the teacher expects the students to perform or

do as a result of instruction.

The technique of writing objectives has evolved over the decades. They

are now also called Specific Learning Outcomes (SLOs).

In the subsequent sections we will be dealing with the mechanics of

writing SLOs.

STEPS FOR STATING SPECIFIC LEARNING OUTCOMES(OR OBJECTIVES)

1. List a representative sample of topics.2. Write a blanket statement.3. Begin each objective with an action verb that specifies

observable performance (e.g. identifies, manages).4. Include enough objectives to describe adequately the

performance of students who have attained the objective.5. Keep the objectives sufficiently comprehensive and global

and not too restricting.

15Assessment of Competence

DAY

ONE

Page 18: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

STATE SLOS AS EXPECTED STUDENT PERFORMANCEStart by writing a blanket statement, an example of which is as follows:“At the end of the session/ unit/ course, the learner will be able to:”l Name the cells involved in acute inflammationl Enlist the drugs that block inflammation and are anti-pyretics.l Differentiate between a normal and an abnormal ECG.l Outline a management plan for a patient with a septic wound and

uncontrolled diabetes mellitus.l Justify investigations required to diagnose a case of high-grade fever in

children under one year of age.

Don’t state them in terms of: l Teacher performance (e.g. teach inflammation)l Learning process (e.g. student learns the concept of immunization)l Course content (e.g. studies the management of enteric fever)l Two objectives (e.g. student knows and understands concepts in

peri-operative management)

USING OBSERVABLE VERBSl Each statement must begin with a verb. However, the verbs are specific

and indicate definite, observable responses; that is, responses that canbe seen and assessed by an observer.

l These verbs state learning objectives in terms of observable studentperformance, e.g. describes, identifies, differentiates, plans.

l Learning objectives should specify the types of reaction the students aresupposed to make to the content and not identify the content itself. So,do not go into too much detail. It is always beneficial to have objectivesthat do not restrict the learning activities. Try and make them global.

OBJECTIVES SHOULD NEITHER BE TOO RESTRICTING NOR TOO VAGUEThe following are too specific and restricting:Describes the phase of isovolumic contraction in a cardiac cycle. Describes the phase of rapid ejection in a cardiac cycle.The following is better:Describes how the heart functions in a normal adult human.

The following is too broad and vague:Deals with cases of endocrine disordesThe following is better:Manages cases of Phaechromocytoma effectively

16Assessment of Competence

DAY

ONE

Page 19: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

TABLE OF SPECIFICATIONSA TABLE OF SPECIFICATIONS is a chart that contains information aboutthe topics of your speciality, the level you want each topic to be taught, i.e.whether you want a particular topic to be at C1, C2 or C3, and the total ofthese topics along with this total expressed in percentage. STEPS IN THE MAKING OF A TABLE OF SPECIFICATION:1. Make a table with the following headings (or columns)

2. First, fill the contents section. write all the topics that you and your

team want.

3. Decide how much ( i.e. relative) importance ( or weightage) you want toassign to each topic

4. In the extreme right column write the % of importance for a particulartopic. For example, in medicine, you may decide to assign 35% to

Hypertension, & 30% to Diabetes Mellitus.

5. Now, continuing with the above example, you have to decide that forHypertension, how much of this 35% you will assign to Cognition,Psychomotor and Affects. Within Cognition, then, you have to decidewhether you want to lay more emphasis on recall ( C1), interpretation(C2) or problem solving (C3). Examples are provided below:

6. Note: It is not possible to have three diffrent levels of Psychomotorsince a person can be at only one level of dexterity at one time. In thecolumn of ‘P’ write either 1, 2 or 3 to indicate the desired outcome.

17Assessment of Competence

DAY

ONE

Contents P A % 1 2 3C o g n i t i o n

Contents P A % 1 2 3C o g n i t i o n

Hypertension, 35

Diabetes

Mellitus 30

Page 20: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

7. Note: For Affective domian use the same principle as that ofPsychomotor domain stated above in point 6.

DECIDING WHICH TOPIC/ UNIT TO GIVE MORE IMPORTANCE TO

This depends on the prevalence of disease in the country. For example, inPakistan if the most prevalent disorder is iron deficiency anemia orpulmonary tuberculosis, then probably most importance should be given toeither ‘Blood’ or ‘Respiratory system’ instead of to CNS or Locomotorsystem..

DECIDING WHETHER ONE SHOULD KEEP MORE OF C1, C2 OR C3

That depends on the level of students you are dealing with. If it is a juniorundergraduate, then there should be relatively more of C1 and C2. If it is asenior undergraduate then the number of C3s should be relatively more.If it is a post graduate student then C3s should be clearly more than theC1s and C2s.

COMMON MISTAKES TO BE AVOIDEDl It is not necessary for ALL the topics to be present in this table. Leave

out the relatively unimportant ones, even if they seem interesting to you.l It is not necessary for every topic to be divided into the two cognitive

domains. It is definitely possible for a topic to have no C2 (e.g. 1st yearMBBS student, topic: CNS) or no C1 (final year MBBS/ resident yearone, topic: CVS).

l Make sure that the total is always given is percentage also. This willhelp you if you want to change the number of items different from thatgiven in the total of the table.

HOW DO WE USE A TABLE OF SPECIFICATION

A table of specifications may be used to decide not just the anatomy of anexam but also the type of instructional strategy that is best suited for aparticular topic. A table of specifications can, thus, help us decide the:

A. Type of instruction that will be used for each sub-topic:l Number of items that should be present in a theory exam paper.l Relative number of topics that should be present in the theory exam.l Cognitive level for assessing each topic and sub-topic l Psychomotor importance that you assign to a unit or sub unit.l Affective importance that you want to assign to a unit or sub unit.

18Assessment of Competence

DAY

ONE

Page 21: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

B. Type of assessment that is applicable for each topic and domain:It doesn’t make sense to assess how well a candidate can perform acertain skill by giving him a written or oral test about it. The best thing todo in such a case is to observe how he performs that skill/ task. So if yoursub unit has some psychomotor component to it, then a good exam wouldinclude not just a written part but also a practical. Similarly if you want toassess if your student has learnt how to introduce himself to the patientand how he deals with the patient then you should have a system ofevaluating his attitude regularly in the OPD or ward.SAMPLE OF COMPLETE TABLE OF SPECIFICATIONS

HOW TO WE USE IT

A table of specifications may be used to decide not just the anatomy of anexam but also the type of instructional strategy that is best suited for aparticular topic. A table of specifications can, thus, help us decide the:

A. Type of instruction that will be used for each sub-topic:l Number of items that should be present in a theory exam paper.l Relative number of topics that should be present in the theory exam.l Cognitive level for assessing each topic and sub-topic l Psychomotor importance that you assign to a unit or sub unit.l Affective importance that you want to assign to a unit or sub unit.

B. Type of assessment that is applicable for each topic and domain:It doesn’t make sense to assess how well a candidate can perform acertain skill by giving him a written or oral test about it. The best thing todo in such a case is to observe how he performs that skill/ task. So if yoursub unit has some psychomotor component to it, then a good exam wouldinclude not just a written part but also a practical. Similarly if you want toassess if your student has learnt how to introduce himself to the patientand how he deals with the patient then you should have a system ofevaluating his attitude regularly in the OPD or ward.

19Assessment of Competence

DAY

ONE

Contents P A % 1 2 3C o g n i t i o n

Hypertension, 16 12 2 30

Diabetes

Mellitus 10 12 8 30

Acute Renal 3 5 2 10

Failure

Tuberculosis 5 6 12 7 30

5 35 41 19 100

Page 22: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

PRINCIPLES OF ASSESSMENT

ASSESSMENT: QUEST FOR IMPROVEMENT

A ssessment is a process of determining the extent to which studentshave achieved the instructional objectives.

The following principles form a general framework within which thisprocess may be viewed:

1. Clarify the purpose of assessment 2. Determine the content that is to be assessed.3. Ensure adequate content sampling by following a table of specifications.4. Select a suitable assessment tool (to fit the purpose).5. Try and use a combination of various assessment tools.6. Use the results to improve learning and training.

Assessment is a means to an end, not an end in itself: once an exam isover, the results should not be stored in a computer or a file but should beused to:

= review the instructions

= review the teaching program

= review teacher performance

= review student performance

= certify students

= upgrade / detain students.

TYPES OF ASSESSMENTFORMATIVE:l This is designed to inform students about the amount he still has to

learn. l It measures the progress made by the learner and enables learning

activities to be adjusted according to the progress made. l It must never be used for certifying purposes since its main aim is

diagnosis of the amount of learning done and needed further.

SUMMATIVE:

l This is designed to protect society from incompetent professionals. l It is traditionally used for placing students in order of merit and justifying

decisions as to whether they should move up to the next level or beawarded a diploma or not.

l It is carried out less frequently than formative assessment and isusually at the end of a unit or period of instruction.

20Assessment of Competence

DAY

ONE

Page 23: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

CHARACTERISTICS OF A GOOD EXAMINATION

An exam is said to be ‘good’ if it has the following characteristics:1. Feasibility2. Objectivity3. Reliability4. Validity

FEASIBILITY:An exam should be cost beneficial and easy to score. It should not be aburden on financial and human resources of the organization.

OBJECTIVITY:An exam is ‘objective’ if the acceptable answer to the questions is definedprior to its administration and any competent person given a key can markit. Objective tests need not be multiple choice only; essays and oral examsmay be objective as well. However the objectivity of the latter two is lowerthan of MCQs.

RELIABILITY:It refers to the lack of error in measurement or consistency ofmeasurement. That is, how consistent test scores or other assessmentresults are from one measurement to another. Reliability is essentiallyconcerned with restricting measurement errors within acceptable limits.

Reliability refers to the results (obtained with an assessment instrument)and not to the instrument itself. So, when you comment about thereliability of a test you are actually talking about its result or score and notabout the test itself. It is important for a test to be reliable because onlythen can we confidently say that this instrument will give us the sameresult if we were to re-administer it to a similar set of students. Reliabilityis usually measured by a computer after the scores have been fed in. Thename of the method is Cronbach’s Alpha Reliability. The value ofCronbach’s alpha varies from 0 to 1. An internationally minimum acceptablevalue is 0.8.

FACTORS IMPROVING RELIABIITY:

Reliability improves with:

= increasing testing time,

= computer marking,

21Assessment of Competence

DAY

ONE

Page 24: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

= good item design,

= item-banking and re-use of proven psychometric quality and

= examiner training.

VALIDITY:Validity is essentially concerned with:

= the content of the assessment ( are the right things being tested?)

= the effect of learning ( are the students learning what we want them tolearn?)

= the assessment methods ( are we using the most appropriate methods forthe purpose?)

In other words, it is the degree to which the test ‘truly’ measures what itis intended to measure. It also includes an evaluation of the adequacy andappropriateness of the uses that are made of the assessment results. So,when we talk of reliability, we are talking about minimizing measurementerrors; when we talk about validity we are talking about appropriateness ofinterpretations made from the results. Validity is a matter of degree; itdoes not exist on an all or none basis. So, there is no such thing as validor invalid. We always refer to a test result as having high validity or lowvalidity. There are many ways of accumulating evidence to support orchallenge the validity of an interpretation or use of an assessment result.Some of these categories are: face validity, content validity, constructvalidity, test-criterion relationship and consequence validity. Validitydoes not have a numeric value. It is a judgement based on the abovementioned evidence.

FACTORS IMPROVING VALIDITYValidity improves with:

= defining the curriculum/ syllabus

= mapping exam content against the curriculum or using a table ofspecifications to develop an assessment plan

= emphasis on essential and important curriculum content

= selecting methods according to objectives

22Assessment of Competence

DAY

ONE

Page 25: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

TYPES OF VALIDITY

Face validity asks, "Does the test ‘appear’ to be suitable for

the level and discipline for which it is intended ?"

Content validity asks, “Is the sampling adequate?"

This is said to be high if the paper setters have followed the

table of specifications.

Construct validity asks, e.g. "Is this test actually assessing

what I want it to assess?"

Predictive validity asks, "How well does performance on

this test predict performance in the future?"

Consequence validity "What is the effect of this

assessment on learning ?

23Assessment of Competence

DAY

ONE

Page 26: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

ESSAY QUESTIONS CHARACTERISTICS

1. FEASIBILITY: high (easy to make and administer, cost effective.)

2. OBJECTIVITY: low (unless a valid key is given to the rater, the marks

will differ widely among the examiners of the same question).

3. RELIABILITY: low. (by Cronbach’s alpha)

4. FACE VALIDITY: medium to high. (They ‘seem’ to measure basic

recall/ knowledge application among medical students of a particular

level.)

5. CONTENT VALIDITY: low. (Will an essay paper really be able to cover

ALL the aspects of a table of specifications given the time constraint?)

6. CONSTRUCT VALIDITY: medium to low. (Considering the type of essay

questions made currently, are we really judging reasoning ability,

knowledge and skills of a subject?)

7. PREDICTIVE VALIDITY: low. (Can an essay paper really predict how

well a student will do in the future in the field of medicine?)

8. CONSEQUENCE VALIDITY: low. (They usually promote rote

memorization of seemingly higher order medical processes such as

pathogenesis, management and diagnoses, so higher order thinking is

not fostered.)

24Assessment of Competence

DAY

ONE

Page 27: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

ESSAY QUESTIONSAn Essay question is a test item in which an examinee composes a

response. It is usually in the form of one or more sentences of a nature

that no single response or pattern of response can be listed as correct and

the accuracy and quality of which can be judged subjectively only by one

skilled or informed in the subject.

Essay questions are one of the oldest formats to assess the knowledge of

a student about the subject. Despite extensive development of new

assessment procedures and criticism on older methods essay type

examinations are still widely used and continue to flourish in many parts

of the world including in United States.

There are two types of essay questions:

== Short essay questions

== Long essay questions

This division depends upon the amount of freedom offered to the

examinee to express his ideas and restriction over the scope of topic.

Essay type question are an appropriate tool for evaluating complex

performance e.g. synthesis of a group of complex concept, summary of a

document, cause analysis.

25Assessment of Competence

DAY

ONE

Page 28: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

SHORT ESSAY QUESTIONS (SEQs)RULES FOR WRITING & ADMINISTERING BETTER SEQs:

SEQs are a good way of measuring an examinee’s ability of higher order

intellectual achievement e.g. synthesis of a group of complex concept,

summary of a document, cause analysis.

They should not be used to test just factual recall

= Define objective clearly

= Formulate well-focused question

= Relate the question to the objective

= Ask question with appropriate terms like summarize, compare, define,

don’t use broader terms like discuss or write etc

= The phrasing of questions should be simple and unambiguous e.g.

terms like ‘describe briefly’ & ‘explain concisely’ should be avoided

= Indicate time limit approximately for each question.

= Indicate marks distribution for each question, especially if it varies

from question to question

= Use of optional question should be avoided because it is difficult to

construct questions of equal difficulty. Furthermore, a better student

may opt for a difficult question without having any added benefit.

= Prepare an expected answer of a question before administering it.

= Pre test the question if possible in field condition. If the question has

already been administered get feed back analysis for record of question

bank.

ADVANTAGES OF SEQs:l SEQs are an important tool of assessing higher cognitive learning

particularly at C3 (problem solving) level.

l Examinees are free to select, relate and present the idea in their own

words.

l Many topics can be examined in a short time (improved sample size)

l SEQs are more reliable & valid than long essays.

l SEQs are precise and clear because they are made according to

well-defined objectives.

l The form of response that is expected is clearly indicated in questions

by using appropriate action words e.g. enumerate, enlist, compare etc.

l Improved precision of objective allows better rating & lessens subjective

bias, although it does not eliminate it completely.

26Assessment of Competence

DAY

ONE

Page 29: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

l The ability of examinees to organize and summarize thoughts is

effectively tested. This leads towards better communication skills.

l Examinees have to think more and write less, thereby relying less on

writing speed than organization of thoughts.

l SEQs give no cues since examinees have to supply responses instead

of selecting them.

l Use of unfair means is not very probable since the time allowed for each

response is very limited.

LIMITATIONS OF SEQs:

l Compared to long essays the reliability of SEQs is high, but generally

their reliability is low if we compare it with that of MCQs. This is because

of difficulty in grading the free response type and inters rater

differences.

l SEQs can be remembered by students and communicated to the next

batch of examinees easily so they are not fit for frequent reuse.

l As the number of SEQs remains limited in any test so few areas are

measured intensely while others are left unattended

l Scoring cost is higher, and time consuming because they can be

checked only by qualified assessors and not by computers or clerks.

l Some people are disadvantaged because of their poor, illegible or slow

writing speed.

27Assessment of Competence

DAY

ONE

Page 30: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

EXAMPLES OF SHORT ESSAY QUESTIONS

A) Question no.1

You are attending patients in the emergency of a tertiary care hospital.

A patient has been brought to you from a village of interior Sind with H/O:

High grade fever –one week.

O/E

Drowsiness present

Dehydration ++

Moderate splenomegaly present.

Name 5 lab investigations and justify how these will help you to reach the

most appropriate diagnosis and subsequent management.

(Your answer should not exceed one page of response sheet)

B) Question no.2

Write the management of a 4-year-old child bitten by a stray cat resulting

in 1/2cm deep wound on face .The cat was hidden in the corner of the

house and attacked the child when he went near it.

(Note: the parents of the child are quite affording)

C) Question no.3

A man, while crossing a road, was hit on the right knee joint by an

automobile. List the probable positive clinical findings you expect if

you examine him two hours after accident in your clinic.

D) Question no.4

A poor woman, working as a domestic servant, is pregnant and severely

anemic. She comes to consult you in a govt. dispensary. You fear that you

will not see this patient again. Discuss the management of this case by

stating not more than five steps.

28Assessment of Competence

DAY

ONE

Page 31: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

SCORING STRATEGIESI. GUIDELINES FOR PREPARING A KEY FOR ESSAY

QUESTIONSA key is a set of instructions for scoring written answers. It defines andsets the minimum factual knowledge/comprehension or ability to solve aproblem, which a candidate must have in order to secure passing marks.

STEPS IN THE PREPARATION OF A SCORING KEY.For knowledge and comprehension questions: Prepare an outline of the expected answer.Assign a score or weightage to each item to be included in the answer.

Example:The following is the key of question A. The pass mark is 60%.

Key:Point m 20 Point n 10 Point o 30 Point p 20 Point q 10 Point r 10

100

The candidate could earn 60% marks by writing

points o + m + n or o + n + p or m + n + p + q or m+n+p+r or any other combination totaling 60%.

If you wish that to pass, the candidate must write one or two facts/ points,which you consider essential, then you should specify these in the key.There are different ways in which this can be done.

Give so much weightage/marks for those particular points that thecandidate cannot get the pass marks without it. In the above examplesuppose you want that point o is mandatory then you can put 50 as theMarks for o and if the candidate misses it or does it very poorly he doesnot get the passing marks. Taking the same example quoted above,suppose 25 marks are deducted for omitting point o. Irrespective of howmany marks the candidate gets in the rest of the question, he/ she cannotpass after losing 30 marks for not having mentioned point o and a furtherdeduction of 25 marks for this act of omission. Thus the total markscandidate losses for missing point o are 55.

29Assessment of Competence

DAY

ONE

Page 32: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

If you want the candidate to mention all the above points then distributethe marks in such a way that they score total 60% or the pass marks forthat question. The rest of the marks could then be for additional points andetter organization.

Note: In certain types of examinations, the candidates are only declaredpass/fall. In such cases the question setter is required to identify theelements that must be present in the answer. These are called the"Essential Points". A candidate must mention all the essential points inorder to pass the question. The additional elements in the answer arecalled

"Optional Points" and are considered only when all the essential pointshave been mentioned.

30Assessment of Competence

DAY

ONE

Page 33: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

II. ANALYTICAL VS HOLISTIC RUBRICS

Combining scores from analytic rubric scaleIf you decide to use an analytic rubric, you may wish to add or averagethe scores from each of the scales to get a total score.

Scoring rubrics may be specific to a particular assignment or they maybe general enough to apply to many different assignments. Usually themore general rubrics prove to be most useful, since they eliminate theneed for constant adaptation to particular assignments and because theyprovide an enduring vision of quality work that can guide both studentsand teachers.

Although it is possible to construct your own rubrics from scratch, manypeople find it simpler to adopt or modify existing rubrics. Examples ofscoring rubrics may be found in education journals and other professionalpublications. These rubrics may be adopted, adapted, or they canprovide ideas and serve as models.

There is no single best rubric for all purposes and many different rubricscould be applied to the same task. The following criteria for evaluatingscoring rubrics are adapted from Herman, Aschbacher and Winters(1992), Arter (1990) and ISBE (1994).

Does the rubric relate to the outcome(s) being measured?The rubric should address all aspects of the outcome(s) being measuredand it should not address anything extraneous. For example, spelling andgrammar might be considered extraneous on a science assessment,unless it is measuring an outcome that deals specifically withcommunication. If the assessment is to address critical thinking skills,does the rubric reflect that?

Does it cover important dimensions of student performance? Do the criteria reflect current conceptions of excellence in the field? Does the rubric reflect what you emphasize in your teaching? Does the highest scale point represent a truly exemplary performance or product? When you evaluate rubrics, you need not be concerned about having acertain number of students score at every point on the scale. It may bethat no student will attain the highest scale point. Nonetheless, it stillmay be worthwhile to have that point on the scale as a standard ofexcellence for which students should strive.

31Assessment of Competence

DAY

ONE

Page 34: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

Are the dimensions or scales well-defined? Is it clear to everyone whateach scale measures?Is there a clear basis for assigning scores at each scale point? Is it clearexactly what a student needs to do to get a score at each scale point?Can you easily differentiate between scale points? An easily understood rubric with clear definitions of each score point is theideal. Conversely, it is usually best to avoid scales that are labeled only atthe highest and lowest points.

Can the rubric be applied consistently by different scorers?Inter-rater reliability depends on how well the scales and scale points aredefined and the extent to which you and your colleagues can arrive atconsensus about how performance should be measured and whatconstitutes good performance.

32Assessment of Competence

DAY

ONE

Page 35: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

ANALYTICAL VS HOLISTIC RUBRICSA rubric with two or more separate scales is called an analytical rubric.This contrasts with a scoring rubric that uses only a single scale that yieldsa global or holistic rating.

Holistic scoring is often more efficient, but analytical scoring systemsgenerally provide more detailed information that may be useful in planningand improving instruction and communicating with students.

Combining scores from analytic rubric scaleIf you decide to use an analytic rubric, you may wish to add or average thescores from each of the scales to get a total score.

Scoring rubrics may be specific to a particular assignment or they may begeneral enough to apply to many different assignments. Usually the moregeneral rubrics prove to be most useful, since they eliminate the need forconstant adaptation to particular assignments and because they providean enduring vision of quality work that can guide both students andteachers.

Although it is possible to construct your own rubrics from scratch, manypeople find it simpler to adopt or modify existing rubrics. Examples ofscoring rubrics may be found in education journals and other professionalpublications. These rubrics may be adopted, adapted, or they can provideideas and serve as models.

There is no single best rubric for all purposes and many different rubricscould be applied to the same task. The following criteria for evaluatingscoring rubrics are adapted from Herman, Aschbacher and Winters (1992),Arter (1990) and ISBE (1994).

Does the rubric relate to the outcome(s) being measured? The rubric should address all aspects of the outcome(s) being measuredand it should not address anything extraneous. If the assessment is toaddress critical thinking skills, does the rubric reflect that?

33Assessment of Competence

DAY

ONE

Page 36: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

Does it cover important dimensions of student performance? Do thecriteria reflect current conceptions of excellence in the field? Does therubric reflect what you emphasize in your teaching? Does the highestscale point represent a truly exemplary performance or product?

When you evaluate rubrics, you need not be concerned about having acertain number of students score at every point on the scale. It may bethat no student will attain the highest scale point. Nonetheless, it still maybe worthwhile to have that point on the scale as a standard of excellencefor which students should strive.

Are the dimensions or scales well-defined? Is it clear to everyone whateach scale measures? Is there a clear basis for assigning scores at eachscale point? Is it clear exactly what a student needs to do to get a scoreat each scale point? Can you easily differentiate between scale points? An easily understood rubric with clear definitions of each score point is theideal. Conversely, it is usually best to avoid scales that are labeled only atthe highest and lowest points.

Can the rubric be applied consistently by different scorers?Inter-rater reliability depends on how well the scales and scale points aredefined and the extent to which you and your colleagues can arrive atconsensus about how performance should be measured and whatconstitutes good performance.

34Assessment of Competence

DAY

ONE

Page 37: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

OPTIONS FOR SELECTING RUBRICS1. Adopt

q Use an existing rubric "as is."

2. Adapt q Modify or combine existing rubrics. q Re-word parts of the rubric. q Drop or change one or more scales of an analytical rubric. q Omit criteria that are not relevant to the outcome you are

measuring. q "Mix and match" scales from different rubrics. q Change the rubric for use at a different grade. q Add a "no-response" category at the bottom of the scale. q Divide a holistic rubric into several scales.

3. Do it yourself q Build your own rubric from scratch about

QUESTIONS TO CONSIDER WHILE MAKING RUBRICS

1. Can the rubric be understood by students and parents? Can it beexplained without technical jargon and in terms that even youngstudents can understand?

2. Is the rubric developmentally appropriate? Does the rubric specifysome score points appropriate for the age of the student?

3. Can the rubric be applied to a variety of questions?The most useful rubrics can be applied to more than one task.

4. Is the rubric fair and free from bias? Does it reflect teachable skillsor does it address variables over which students and educatorshave no control, such as the student's culture, gender or homeresources? Does the rubric reward or penalize students based onskills unrelated to the outcome being measured? Have all studentshad an equal opportunity to learn the content and skills addressedin the rubric? Is the rubric appropriate for the conditions underwhich the task was completed?

5. Is the rubric useful, feasible, manageable and practical? Does therubric make sense to you? Will it provide the kind of informationyou need and can use effectively? Does the rubric have areasonable number of scales and score points?

35Assessment of Competence

DAY

ONE

Page 38: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

It is difficult to rate student work on many scales at once and it may behard for students to assimilate data from many score scales. Using longscales is often more time-consuming than using shorter ones.

It's impossible to know for sure how well a rubric will work until you andyour colleagues actually try it out on some actual examples of studentwork. You may want or need to pilot test several rubrics on the samestudent products or performances (it's easier if you can tape theperformances) to see which one works best for you. It's also important tosee whether you and others can learn to apply the rubric consistentlyenough so that you can generally agree on what score to assign to a pieceof student work. This consistency or reliability is crucial if the assessmentis to yield valid, meaningful data.

One method for assessing reliability involves having each raterindependently score several examples of work produced by high-achieving,average, and low-achieving students, and note the extent to which theraters assign the same score to a piece of work. If there is no consensus,having the raters explain and discuss their scores can be instrumental inclarifying the rubric and fostering uniform scoring.

SUMMARY:

Things to consider when selecting a rubric

l Does the rubric relate to the outcome(s) being measured? Does it

address anything extraneous?

l Does the rubric cover important dimensions of student performance?

l Do the criteria reflect current conceptions of "excellence" in the field?

l Are the categories or scales well-defined?

l Is there a clear basis for assigning scores at each scale point?

l Can the rubric be applied consistently by different scorers?

l Can the rubric be understood by students and parents?

l Is the rubric developmentally appropriate?

l Can the rubric be applied to a variety of tasks?

l Is the rubric fair and free from bias?

l Is the rubric useful, feasible, manageable and practical?

36Assessment of Competence

DAY

ONE

Page 39: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

QUALITY ASSURANCE IN RUBRICS:

In everyday life we usually try to get the facts before we make a decision.The greater the amount of accurate information we can gather, the moreconfident we can be that we are making the right decision. The same thingholds true in the classroom when we assess students to get informationthat's used to decide what they know and can do, to diagnose their needs,to decide if we need to teach something in a different way, to planinstruction, to assign grades, to communicate with parents, and to guideus in the development of school improvement plans. The consequencesof making a poor decision can be very serious, so it's important to makesure that we're basing our decisions on an adequate amount of soundinformation. Or to put it another way we need assessment results that areboth reliable and valid for our particular purpose.

An assessment is reliable if it yields results that are accurate and stable.In order for a performance assessment to be reliable, it should beadministered and scored in a consistent way for all the students who takethe assessment.

An assessment is valid for a particular purpose if it in fact measures whatit was intended to measure. An assessment of a learning outcome is validto the extent that scores truly measure that outcome and are not affectedby anything irrelevant to the outcome.

Some important aspects of validity are content coverage, generalizabilityand fairness. The assessments for a given outcome should be alignedwith the both the outcome and instruction and, when taken together,should cover all important aspects of the outcome. The assessmentsshould address the higher-order thinking skills specified in the outcome.The tasks used should have answers or solutions that can't be memorized,but which, instead, call on the student to apply knowledge and skills to anew situation.

Assessment results are generalizable [to the extent that] if there isevidence that scores on one assessment can predict how well studentsperform on another assessment of the same outcome.

Once you decide on a rubric, the best way to promote reliable scoring is tohave well-trained scorers who thoroughly understand the rubric and whoperiodically score the same samples of student work to ensure that theyare maintaining uniform scoring.

Another way to increase reliability is to try hard to stick to the rubric as youscore student work. Not only will this increase reliability and validity, but

37Assessment of Competence

DAY

ONE

Page 40: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

it's only fair that the agreed-upon rubric that's shared with students andparents should be what is actually used to rate student work.Nonetheless, human beings making subjective judgments mayunintentionally rate students based on things that aren't in the rubric at all.The conscientious scorer will frequently monitor his or her thinking toprevent extraneous factors from creeping into the assessment process.The box below contains a list of some extraneous factors to watch out for.

After you decide upon a rubric, you might find it helpful to reformat it in away that makes it more easily usable by you and your students. Forexample, an analytical scale may be easier to use or more meaningful if itis typed or summarized in tabular form.

38Assessment of Competence

DAY

ONE

Page 41: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

PROBLEMS AND PITFALLS ENCOUNTERED BY SCORERS

Positive-Negative Leniency Error: The scorer tends to be too hard or tooeasy on everyon.Trait Error: The scorer tends to be too hard or too easy on a given trait,criterion or scale.Appearance: The scorer thinks more about how the paper or project looksthan about the quality.Length: Is longer better? Not necessarily.Fatigue: Everybody gets tired.Repetition Factor: This paper is just like the last 50. Order Effects: If you've just read 10 bad papers, an average one may startto look like Shakespeare by comparison.Personality Clash: It's tougher if you don't like the topic or the student'spoint of view.Skimming: Doesn't the first paragraph pretty well tell the story?(Hint: No.)Error of Central Tendency: Using an odd-numbered scoring scale? Bewarethe dreaded "mid-point dumping ground."Self-Scoring: Are you a perceptive reader? Be sure what you're scoring isthe writer's work--not your own skill.Discomfort in Making Judgments: Remember that you are rating thepaper, product or performance, not the student. This is just oneperformance assessment--not overall ability.The Sympathy Score: "The student was really trying..." "...seems to besuch a nice kid," "...chose a hard topic," "...had a tough day," etc.--Adapted from Culham and Spandel (1993)

REFERENCES:l Arter, Judith. Per formance Rubric Evaluation Form (Metarubric).

Portland, OR: Northwest Regional Educational Laboratory, September1990.

l Culham, Ruth and Spandel, Vicki. Problems and Pitfalls Encountered byRaters. Developed at the Northwest Regional Educational Laboratoryfor the Oregon Department of Education, 1993.

l Herman, Joan, Aschbacher, Pamela, and Winters, Lynn. A PracticalGuide to Alternative Assessment. Alexandria, VA: Association forSupervision and Curriculum Development, 1992

39Assessment of Competence

DAY

ONE

Page 42: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

40Assessment of Competence

Exc

elle

nt

(10)

The

writ

er p

rovi

des

high

ly lo

gica

lar

gum

ents

. B

oth

(pos

itive

and

neg

ativ

e)as

pect

s ar

e pr

esen

ted

very

nic

ely.

The

writ

erha

s ve

ry c

lear

lypr

esen

ted

his/

her

ow

nst

ance

on

the

topi

c an

dha

s de

fend

ed it

suita

bly.

Pro

per

refe

renc

es a

re p

rovi

ded

for

the

argu

men

ts.

The

info

rmat

ion

pres

ente

d m

atch

esex

actly

with

the

requ

irem

ents

. It

isev

iden

t th

at

refe

renc

es a

re f

rom

varie

d so

urce

s an

din

clud

e a

good

num

ber

of n

ew o

nes

Typi

cally

cle

arbe

ginn

ings

, m

iddl

es,

and

ends

.N

o di

gres

sion

s ar

epr

esen

t.Tr

ansi

tions

are

sm

ooth

and

logi

cal.

Mec

hani

cal e

rror

s ar

eab

sent

.Id

eas

are

wel

lor

gani

zed

and

com

mun

icat

ed c

lear

lyan

d co

ncis

ely.

The

lear

ner

has

deve

lope

d a

high

lylo

gica

l pie

ce o

f w

ork

base

d on

cle

arly

defin

ed p

rinci

ples

/th

eory

. The

re s

eem

s to

be n

o co

nfus

ion

in t

hele

arne

r’s m

ind

as t

oho

w t

o us

e th

e th

eory

/pr

inci

ples

tha

t w

ere

prov

ided

. The

wor

kde

velo

ped

is h

ighl

yus

eabl

e.

Go

od

(08)

The

writ

er p

rovi

des

logi

cal a

rgum

ents

. B

oth

(pos

itive

and

neg

ativ

e)as

pect

s pr

esen

ted

are

bala

nced

. The

writ

erha

s tir

ed t

o pr

esen

t hi

s/he

r ow

n st

ance

on

the

topi

c an

d ha

s de

fend

edit

at m

ost

plac

es.

Arg

umen

ts a

resu

ppor

ted

by p

rope

rre

fere

nces

.

The

info

rmat

ion

pres

ente

d m

atch

es t

here

quire

men

ts.

It is

evid

ent

that

re

fere

nces

are

fro

mva

ried

sour

ces

and

incl

ude

a go

od n

umbe

rof

new

one

s

Gen

eral

ly w

ell

orga

nize

d ac

cord

ing

tode

finite

pla

ns.

Typi

cally

cle

arbe

ginn

ings

and

end

s.M

ost

tran

sitio

ns a

resm

ooth

and

logi

cal.

Mec

hani

cal e

rror

s do

not

conf

use

read

er,

but

in e

ach

com

posi

tion

ther

e m

ay b

e se

vera

lm

inor

err

ors

or o

ne o

rtw

o m

ajor

err

ors.

The

lear

ner

has

deve

lope

d a

cred

ible

piec

e of

wor

k ba

sed

oncl

ear

cut

prin

cipl

es /

theo

ry. T

here

see

ms

tobe

alm

ost

no c

onfu

sion

in t

he le

arne

r’s m

ind

asto

how

to

use

the

theo

ry/

prin

cipl

es t

hat

wer

e pr

ovid

ed. T

hew

ork

deve

lope

d is

high

ly u

seab

le.

Lev

el o

f ac

hie

vem

ent

Arg

um

enta

tio

n/

Just

ific

atio

ns

Co

nte

nt

Exp

ress

ion

(sty

le o

f w

riti

ng

)P

ract

ical

Ap

plic

atio

n

AN

ALY

TIC

AL

RU

BR

IC E

XA

MP

LE

Page 43: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

41Assessment of Competence

Ad

equ

ate

(06)

The

writ

er p

rovi

des

logi

cal a

rgum

ents

.A

dequ

ate

bala

nce

ism

issi

ng in

onl

y a

few

case

s. T

he w

riter

has

trie

d to

pre

sent

his

/ he

row

n st

ance

on

the

topi

can

d ha

s de

fend

ed it

at

mos

t pl

aces

. S

uita

ble

refe

renc

es a

re p

rovi

ded

for

the

argu

men

ts in

mos

t ca

ses.

The

info

rmat

ion

pres

ente

d is

mos

tlyac

cord

ing

to w

hat

has

been

ask

ed. T

henu

mbe

r fo

r re

fere

nces

incl

uded

is b

arel

ym

inim

al t

o su

ppor

t th

ear

gum

ents

. La

test

refe

renc

es a

re in

suffi

cien

t nu

mbe

rs.

End

ings

may

som

etim

es b

e aw

kwar

dor

abr

upt.

Tran

sitio

ns a

re t

ypic

ally

logi

cal b

ut m

ay o

noc

casi

on la

ck d

epth

and/

or d

irect

rel

evan

ce.

Dig

ress

ions

may

be

foun

d he

re a

nd t

here

but

they

do

not

ham

per

the

flow

of

text

.

The

writ

er is

unc

lear

at

som

e pl

aces

reg

ardi

ngw

hat

prin

cipl

es/

theo

ryto

app

ly. T

here

is s

ome

conf

usio

n in

the

lear

ner’s

min

d as

to

how

to

use

the

theo

ry/

prin

cipl

es t

hat

wer

epr

ovid

ed. T

he w

ork

deve

lope

d is

vag

ue a

tso

me

plac

es o

nly.

At

som

e pl

aces

lack

s flo

wof

pro

cess

/ th

ough

t.M

ajor

ity o

f th

e id

eas/

prop

ositi

ons

are

usea

ble.

Inad

equ

ate

(04)

The

writ

er p

rovi

des

logi

cal a

rgum

ents

.B

alan

ce is

mis

sing

inso

me

case

s. T

he w

riter

pres

ents

his

/ he

r ow

nst

ance

on

the

topi

c bu

tfa

ils t

o de

fend

itad

equa

tely

. R

efer

ence

sar

e no

t pr

ovid

ed f

or t

hear

gum

ents

in m

ost

case

s.

The

info

rmat

ion

pres

ente

d is

at

plac

esac

cord

ing

to w

hat

has

been

ask

ed. T

henu

mbe

r fo

r re

fere

nces

incl

uded

is in

adeq

uate

to s

uppo

rt t

hear

gum

ents

. La

test

refe

renc

es a

rein

suffi

cien

t in

num

ber.

Pos

sibl

e di

gres

sion

s or

elab

orat

ions

are

conf

usin

g to

rea

der.

Som

e tr

ansi

tions

are

logi

cal.

Beg

inni

ngs

and

endi

ngs

may

be

awkw

ard

or a

brup

t.K

ey e

lem

ents

may

be

unev

enly

dev

elop

ed o

rom

itted

.M

echa

nica

l err

ors,

som

e m

inor

, so

me

maj

or,

whi

ch m

ay o

noc

casi

on c

onfu

sere

ader

.

The

writ

er is

unc

lear

at

mos

t pl

aces

reg

ardi

ngw

hat

prin

cipl

es/

theo

ryto

app

ly. T

here

isco

nfus

ion

in t

hele

arne

r’s m

ind

as t

oho

w t

o us

e th

e th

eory

/pr

inci

ples

tha

t w

ere

prov

ided

. The

wor

kde

velo

ped

is v

ague

at

mos

t pl

aces

. At

mos

tpl

aces

lack

s flo

w o

fpr

oces

s/ t

houg

ht.

Maj

ority

of

the

idea

s/pr

opos

ition

s ar

eus

eabl

e.

Lev

el o

f ac

hie

vem

ent

Arg

um

enta

tio

n/

Just

ific

atio

ns

Co

nte

nt

Exp

ress

ion

(sty

le o

f w

riti

ng

)P

ract

ical

Ap

plic

atio

n

Page 44: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

42Assessment of Competence

Po

or

(02)

The

writ

er f

ails

to

prov

ide

any

argu

men

tat

all.

The

writ

er d

oes

not

pres

ent

his/

her

own

stan

ce o

n th

e to

pic

and

fails

to

defe

nd it

.R

efer

ence

s ar

e no

tpr

ovid

ed f

or t

hear

gum

ents

.

Lack

of

nece

ssar

yin

form

atio

n.To

pic

may

be

clea

r bu

tno

ove

rall

orga

niza

tiona

l pla

nsex

ist.

The

re a

re m

any

digr

essi

ons

or o

ver-

elab

orat

ions

or

little

deve

lopm

ent

alto

geth

er.

The

re is

littl

e se

nse

ofbe

ginn

ings

or

endi

ngs.

Man

y m

echa

nica

ler

rors

inte

rfer

e w

ithun

ders

tand

ing.

Inco

mpl

ete

sent

ence

sar

e pr

esen

t.

Inco

mpl

ete

sent

ence

sar

e pr

esen

t.T

here

see

ms

to b

e to

tal

lack

of

clar

ity t

o th

ere

ader

as

to w

hat

prin

cipl

es/

theo

ry h

asbe

en a

pplie

d. T

here

seem

s to

be

grea

tco

nfus

ion

in t

hele

arne

r’s m

ind

as t

oho

w t

o us

e th

e th

eory

/pr

inci

ples

tha

t w

ere

prov

ided

. The

wor

kde

velo

ped

is v

ague

,la

cks

flow

of

proc

ess/

thou

ght

and

does

not

seem

to

be u

seab

le.

Page 45: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

43Assessment of Competence

Exc

elle

nt

(10)

Go

od

(08)

Lev

el o

f ac

hie

vem

ent

BLA

NK

RU

BR

IC T

EM

PLA

TE F

OR

PR

ACTI

CE

Page 46: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

44Assessment of Competence

Ad

equ

ate

(06)

nad

equ

ate

(04)

Lev

el o

f ac

hie

vem

ent

Page 47: Yecf[j[dY[ - CPSPelogbook.cpsp.edu.pk/elogbook/supervisors_workshop... · 2011-03-16 · Contents Page No. 1 - 2 Introduction 3 Overview 5 Competence 6 Systems approach 9 Taxonomy

45Assessment of Competence

Po

or

(02)