Download - Assessment Q & A, Examples, and Tools · Use Bloom‘s Taxonomy and active verbs (create, analyze, demonstrate, etc.) Be written in the positive instead of the negative Reflect measurable

Transcript

Assessment Q & A, Examples, and Tools

What is a student learning outcome (SLO)?

Student Learning Outcome (SLO): An SLO identifies the measurable knowledge, skills, behaviors, or attitudes of the learner as the result of engaging in a learning activity or program. Typically, SLOs are composed with the stem, “The student will…”.

What is Assessment?

The systematic collection and analysis of information to improve student learning and program viability. Assessment is “…the process of gathering evidence to make inferences about…how students are progressing toward specific goals” (National Standards, quote form Pennington, 2001,p. 206)

What is Value-Added Assessment? Value-Added is an analytical strategy to determine the degree to which students change from the beginning to the end of a program. Astin (1985) referred to this type of change as talent development.

Aren’t the SLOs (Student Learning Outcomes) essentially the same thing as the SOLs (Standards of Learning) that are creating havoc in the public schools?

No, actually they are quite different. The SOLs really focus on student assessment, whereas our SLOs are meant to be course/program assessment. The goal of the SOLs is to evaluate individual student achievement in a state imposed curriculum and determine whether they are ready to go on to the next grade. SOLs are also used to evaluate whether teachers or schools are successful in getting al their students where they need to be to precede to the next level.

Our SLOs are of our own choosing, within the parameters of the VCCS course/program guide. They are the specific, measurable skills our faculty have stated they want students to achieve in a particular course/program. It is our job to develop assessment tools that measure these, in a way that does not hinder individual teaching styles or methods and promotes sharing of best practice and good ideas.

Why aren’t grades enough?

When faced with the news that it‘s your discipline‘s turn for outcomes assessment, it is tempting

to ask why you can‘t just look at final grades to determine whether a course is successful.

Although counting letter grades is easy, it provides neither consistent nor meaningful

information about student success in a multi-section course.

2

In outcomes assessment, the terms ―scoring‖ and ―grading‖ have different meanings. Scoring

refers to the process of marking an assessment instrument to get data about how well the course

has done at achieving its outcomes. Grading is the process of marking an assessment instrument

for the purpose of assigning a student a grade for the course. Scoring needs to be done

consistently across all sections; grading can be done differently in each section if instructors

desire. In no way, does the outcome assessment scoring process infringe on an instructor‘s

grading.

Unless every instructor teaching a particular course assigns final course grades in exactly the

same way (same assignments, same exams, same weights, same grading approach), you cannot

be confident that one section‘s A is the same as another section‘s A. More significantly, final

grades are an aggregate assessment of a student‘s entire work for the course, often including

attendance and class participation. Consequently, looking at a distribution of grades will provide

little, if any, useful information about the degree to which students are learning those things that

instructors deem most important in the course.

Course Grades versus Course Assessment

Course grades do not provide the same insight that a course assessment does.

Grades give a global evaluation but do not provide sufficiently detailed information about

which course outcomes students are mastering well and which are giving them trouble.

Course grades alone don‘t stimulate faculty discussions about how to improve student

learning of particular course outcomes.

Grades sometimes are based on more than mastery of course content; for example,

participation, attendance, bonus points.

Grading standards often vary widely among different instructors and do not indicate the

same degree of mastery of course outcomes.

Grade inflation (easy tests, generous grading, extra-credit bonuses) sometimes presents a

misleading indicator of student mastery of course outcomes.

The list below shows additional differences between assessment versus grades: Assessment Grades

Formative Summative

Formative refers to the formation of a concept or item whereas summative refers to an ―adding-up‖ or summary

stage. Assessments usually occur in mid-progress when corrections can be made. Grades are usually recorded at the

end of a project or class in order to summarize academic quality.

Diagnostic Final

Non-Judgmental Evaluative

Assessment is non-judgmental in the sense that it focuses on learning, which is the outcome of many influences,

including teaching style, student motivation, time on task, study intensity, and background knowledge. Therefore, no

one element can be reasonably singled out for praise or blame for a particular learning outcome. In contrast, grades

carry evaluative weight as to the worthiness of student achievement and are applied, for good or ill, directly to them.

Private Administrative

Assessment tends to be used in private and become public only under the assessor‘s control. Grades, while not truly

public, are part of the administrative record available throughout an educational institution.

Often Anonymous Identified

Assessment is almost always collected in anonymous fashion and the results are released in the aggregate. Grades

3

are identified with specific students.

Partial Integrative

To use a metaphor from the calculus, assessment more resembles a partial derivative whereas grades are more

recognizable as in integrative process.

Specific Holistic

Assessment tends to look at specific parts of the learning environment. Grades are holistic in the sense that they

record academic achievement for a whole project. Final grades, of course, can reduce academic achievement for an

entire semester to a single mark.

Mainly Subtext Mostly Text

The text of a course is its disciplinary content; grades tend to focus on that. The subtext of a course involves the

transferrable baccalaureate skills, such as critical thinking, creative thinking, writing, and analysis. For example, the

―text‖ of a course in anatomy and physiology includes the names of bones and functions or muscles. The ―subtext‖

of such a course might include scientific thinking, problem solving, and memory improvement. Grades tend to focus

on text; assessment tends to emphasize subtext.

Suggestive Rigorous

Assessment findings tend to be suggestive and have pedagogical significance. That is, assessment findings shift

pedagogy for reasons that need not be justified statistically, but can be justified when even one student learns better.

In contrast, grades are recorded in a rigorous manner that does have statistical significance.

Usually Goal-Directed Usually Content-Driven

As with text and subtext mentioned above, grades tend to reflect student control of disciplinary course content

whereas assessment usually aims at the goals for all baccalaureate students, such as synthetic thinking and esthetic

appreciation.

How do you write SLOs for a course or program?

A student learning outcome statement needs to specify who is to perform (student), what action

they are to take, and some result that must come from their action. A student learning outcomes

(SLOs) for a course/program should:

Be written in terms of what the student/graduate will be able to do at the end of the

course/program

Be limited to 2-5 outcomes

Keep them short and simple (KISS)

Make them specific, measurable, attainable, realistic, and timely (S.M.A.R.T)

Establish a target performance level for success (i.e. 70% will …)

Keep the assessment process manageable and meaningful (M&M)

Use Bloom‘s Taxonomy and active verbs (create, analyze, demonstrate, etc.)

Be written in the positive instead of the negative

Reflect measurable standards (benchmarks) or reflect the basic knowledge and skills that

the student will be held accountable for

Reflect a combination of higher order thinking skills and supporting or enabling skills

What is the difference between course assessment vs. class assessment?

Course assessment measures the student learning that takes place in ALL sections of the course

for the entire college. It is not to be confused with assessment of instructors or employment

evaluation.

A course assessment consists of all the classes (sections) being taught; for example, ENG 111. A

class assessment is one section of a course, ENG 111-51A or ENG 111-61B.

4

How important is it to design Course Assessment processes to include students from all locations that a course is taught? Inclusion of students from all locations that a course is taught is crucial to the process of Course Assessment. This means that if a course is taught through dual enrollment or is web-based, at both campuses, or off-site, the mechanism for gathering data for assessment of that course must be designed to reasonably include students regardless of the location or delivery-method of the specific section in which they were enrolled. Therefore, if an end-of-course or beginning-of-course activity is developed for the purposes of course assessment, the activity must take place in all sections of the applicable course during the semester that data is being collected. The practical logistics of this requirement may influence some choices of such activities. It is important when designing course assessment to consider the logistics of gathering college-wide data. Must data towards Course Assessment be limited to data collection at the end of the semester? Absolutely not! In fact, some of the most valuable data can be captured in creative ways from students who completed the course being assessed during the previous semester. For example, students who are beginning Chemistry 112 could be given the first-day-of-class assessment covering the course objectives from Chemistry 111. The CHM 112 instructor can review the results to get an idea of what the students have retained, and then pas those assessment forms to the individual responsible for collating the CHM 111 assessment data. This strategy obviously can work for any two-course required-sequence. Finally, instructors may find that they can make arrangements with colleagues, either in their same discipline or across disciplines, to collaborate in activities for course assessment. An example of this could be in a nursing course, where it might be appropriate to include either a formative or a summative assessment of students’ knowledge of infant to adolescent developmental psychology. Aggregate student performance information on this assessment could then be turned over to the psychology faculty for use in their assessment of the Developmental Psychology course. What is the difference between course assessment vs. program assessment? Whereas course assessment focuses on the question of “how can the course be strengthened based on how well students are mastering course objectives?”, program assessment focuses on student learning outcomes for the program as a whole, as well as productivity measures related to the viability and effectiveness of a degree or certificate program. Annual student learning outcome assessments are done in all programs and a program review is evaluated using a 5-year cycle.

5

What is the difference between direct and indirect assessment? Direct Assessment Methods: Direct assessment methods give instructors measurable data to

study. Some examples are written exams, oral exams, performance assessments, standardized

tests, licensure exams, oral presentations, projects, demonstrations, case studies, simulations,

portfolios and, juried activities with outside panels.

Indirect Assessment Methods: Indirect assessment methods provide extra information that may

be used to make changes. Examples include questionnaires, interviews, focus groups, employer

satisfaction studies, observations of advisory boards, and job/transfer school placement data.

Example Course Syllabus Using Student Learning Outcomes Example Revision to Course Syllabus to Show Student Learning Outcomes

Old Course Objectives New Course SLO Objectives

Students will …

To introduce students to PDCCC and Virginia Community College System policies and procedures

To promote support services available to students

To familiarize students with skills necessary for successful college adjustment

To apply the organization skills necessary for college success (time management, stress management, note-taking, etc.)

To familiarize students with the Learning Resources Center web page and information literacy

To apply effective study skills and memory techniques

To explore career possibilities and formulate a tentative career plan

To understand the curricular planning process

To understand the college transfer process

To teach interpersonal communication skills

To provide essential facts concerning AIDS, alcohol and substance abuse

To familiarize students with PDCCC’s Web page and e-PDCCC

Possess knowledge of Paul D. Camp Community College’s policies, procedures, and resources

Demonstrate necessary survival skills for college success (critical thinking, financial planning, memory techniques, note-taking, study skills, and time management)

Demonstrate ability to use a computer to access the Internet, the college website, the blackboard site, and send and reply to email

Demonstrate communication skills (oral and written)

Demonstrate knowledge of personal development areas, such as, essential facts concerning AIDS, alcohol, and substance abuse

Demonstrate an understanding of the career planning process

6

Creating Student Learning Outcomes

To model writing student learning objectives in a straightforward and non-threatening manner,

the following chart uses levels of understanding from Bloom‘s Taxonomy, combines them with

action verbs, and provides examples for a variety of disciplines.

Example: Student Learning Objectives (SLO) Using Blooms Taxonomy

If I want to measure

knowledge outcomes,

I might write…

The student/graduate will…

– Describe the basic components of empirical research.

– Give examples of major themes or styles in music, art, or theatre.

– Recognize in complex text local, rhetorical, and metaphorical patterns.

If I want to measure

comprehension

outcomes, I might

write…

The student/graduate will…

– Correctly classify a variety of plant specimens.

– Explain the scientific method of inquiry.

– Summarize the important intellectual, historical, and cultural traditions in music, art,

or theatre from the renaissance to modern times.

If I want to measure

application outcomes,

I might write…

The student/graduate will…

– Demonstrate in the laboratory a working knowledge of lab safety procedures.

– Apply oral communication principles in making a speech.

– Compute the area of a room.

– Use editing symbols and printers‘ marks.

If I want to measure

analysis outcomes,

I might write…

The student/graduate will…

– Distinguish between primary and secondary literature.

– Diagram a sentence.

– Listen to others and analyze their presentations.

– Differentiate between historical facts and trivia.

If I want to measure

synthesis outcomes,

I might write…

The student/graduate will…

– Revise faulty copy for a news story.

– Formulate hypothesis to guide a research study.

– Create a poem, painting, design for a building.

If I want to measure

evaluation outcomes,

I might write…

The student/graduate will…

– Compare art forms of two diverse cultures.

– Critically assess an oral presentation.

– State traditional and personal criteria for evaluating works of art.

– Draw conclusions from experimental results.

7

Bloom’s Taxonomy: Action Verb List (Partial List) Students/Graduates will …

Cognitive Knowledge will be able to acquire will be able to collect will be able to define will be able to distinguish

will be able to examine will be able to identify

will be able to label will be able to list will be able to name will be able to quote will be able to recall

will be able to recognize will be able to show will be able to tabulate will be able to tell

Comprehension will be able to associate will be able to change will be able to conclude will be able to contrast will be able to

demonstrate will be able to describe will be able to determine will be able to

differentiate will be able to discuss will be able to

distinguish will be able to draw

will be able to estimate will be able to explain will be able to extend will be able to

extrapolate will be able to fill in will be able to give in

own words will be able to illustrate will be able to infer will be able to

interpolate will be able to interpret will be able to make

will be able to predict will be able to prepare will be able to read will be able to rearrange will be able to reorder will be able to rephrase will be able to represent will be able to restate will be able to

summarize will be able to transform will be able to translate

Application will be able to apply will be able to calculate

will be able to change will be able to choose

will be able to classify will be able to complete

8

will be able to demonstrate

will be able to develop will be able to discover will be able to employ will be able to examine

will be able to experiment

will be able to generalize will be able to illustrate will be able to modify will be able to organize

will be able to relate will be able to

restructure will be able to show will be able to transfer will be able to use

Analysis will be able to analyze will be able to arrange will be able to categorize will be able to classify will be able to compare will be able to connect will be able to contrast will be able to deduce

will be able to detect will be able to

discriminate

will be able to distinguish

will be able to divide will be able to explain will be able to explain

will be able to identify will be able to infer will be able to order will be able to recognize will be able to select will be able to separate

Synthesis will be able to classify will be able to combine will be able to compose will be able to constitute will be able to create will be able to deduce will be able to derive will be able to design will be able to develop will be able to document will be able to formulate

will be able to generalize will be able to integrate will be able to invent will be able to modify will be able to organize will be able to originate will be able to plan will be able to prepare will be able to produce will be able to propose will be able to rearrange

will be able to relate will be able to rewrite will be able to specify will be able to substitute will be able to

synthesize will be able to tell will be able to transmit will be able to write

Cognitive (Continued) Evaluation will be able to apprise will be able to argue will be able to assess will be able to compare will be able to conclude will be able to consider will be able to contrast will be able to convince will be able to decide

will be able to decide will be able to

discriminate will be able to explain will be able to grade will be able to judge will be able to measure will be able to rank will be able to

recommend will be able to select will be able to

standardize will be able to

summarize will be able to support will be able to test will be able to validate

Affective Receiving will choose to accept will choose to accumulate

will choose to combine will choose to control will choose to

differentiate will choose to listen (for) will choose to posturally

respond to will choose to select

will choose to separate will choose to set apart will choose to share

9

Responding will choose to acclaim will choose to applaud will choose to approve will choose to augment will choose to commend

will choose to comply (with)

will choose to discuss will choose to follow will choose to play

will choose to practice will choose to spend

leisure time in will choose to volunteer

Valuing will choose to assist will choose to debate will choose to deny will choose to help will choose to increase

will choose to increase numbers of

will choose to protest will choose to relinquish will choose to specify

will choose to subsidize will choose to support will choose to argue

Organization will choose to abstract will choose to balance will choose to compare

will choose to define will choose to discuss will choose to formulate

will choose to organize will choose to theorize

(on) Characterization by Value will choose to avoid will choose to be rated

high by peers in will choose to be rated

high by

will choose to be rated high by superiors in

will choose to change will choose to complete will choose to manage

will choose to require will choose to resist will choose to resolve will choose to revise

10

Watch Out for Verbs that are not Measurable In order for an objective to give maximum structure to instruction, it should be free of vague or ambiguous words or phrases. The following lists notoriously ambiguous words or phrases which should be avoided so that the intended outcome is concise and explicit.

48 Bad Words or Phrases Avoid using verbs that are difficult to measure objectively. Following are examples of verbs difficult to assess, and should be used with caution:

Appreciate Comprehend Experience Realize

Be aware Cover Have faith in Recognize

Memorize Enjoy Internalize Study

Conceptualize Familiarize Know Understand

Believe Feel Learn Values

Hear Capacity Intelligence See

Think Listen Self-Actualize Depth

Be comfortable with … Be acquainted with … Grasp significance of… Perceive

Gain knowledge of … Appreciation for … Acquainted with … Attitude of …

Adjusted to… Awareness of… Capable of … Comprehension of …

Cognizant of … Enjoyment of … Conscious of … Feeling for …

Familiar with … Interest in … Knowledge of … Self-Confident in …

11

Evaluation Method to Measure Outcomes Method A

Writing Effective and Measurable Objectives: The A-B-C-D Model

Element Description Example

A = Audience Who is performing the action? Learning objectives are always stated in terms of student outcomes.

Following completion of the Science program, the student should be able to plot a quadratic equation using a graphing calculator in two minutes or less.

B = Behavior What will the student be able to do? Use an action verb that describes an accomplishment that is measurable. Be specific. Choose a verb that expresses the skill and cognitive level that you want the student to exhibit. (See Bloom’s Taxonomy)

Following completion of the Science program, the student should be able to plot a quadratic equation using a graphing calculator in two minutes or less.

C = Condition Give the conditions under which the performance will occur. Be specific. Conditions should communicate the situation, tools, references, or aids.

Following completion of the Science program, the student should be able to plot a quadratic equation using a graphing calculator in two minutes or less.

D = Degree Describe the minimum criteria for acceptable student performance. Define expectations regarding accuracy, quality, and speed. Be specific

Following completion of the Science program, the student should be able to plot a quadratic equation using a graphing calculator in two minutes or less.

Note: Current educational practices recommend that the audience (student) and the behavior be connected with the terms “should be able to” since faculty cannot promise that everyone will accomplish the stated objective.

A-B-C-D Model

Writing objectives isn’t creative writing: Just follow a formula!

Given [Conditions] the [Audience] will [Behavior] by [Degree].

[Audience] will [Behavior] to [Standard] when provided [Conditions].

12

Method B

S.M.A.R.T Objectives to Generate Outcomes

Specific A specific objective has a much greater chance of being accomplished

than a general goal (Who, What, Where, When, Which and Why)

General Goal – This year I am going to get into shape.

Specific Objective – This year I am going to join a health club and

workout 3 days a week.

Measurable Establish concrete criteria for measuring progress toward the attainment of

each objective you set

Stay on track, reach target dates and experience achievement

How much? How many? How will I know when it is accomplished?

Attainable When you identify objectives that are most important to you, you begin to

figure out ways you can make them come true.

You develop attitudes, abilities, skills, and financial capacity to reach

them.

You can attain most any goal you set when you plan your steps WISELY

and establish a time frame that allows you to carry out those steps.

Realistic To be realistic, an objective must represent something towards which you

are both WILLING and ABLE to work.

Your objective is probably realistic if you truly BELIEVE that it can be

accomplished.

Timely An objective should be grounded within a timeframe. With no timeframe

tied to it there‘s no sense of urgency.

When you set a timeframe, then you have set your unconscious mind into

motion to begin working on the goal.

13

Methods of Assessment

“How do I assess thee, let me count the ways.” Writing Performing Creating/Developing

Essay Demonstration Video

Report Role play Poster

Journal/reflective writing Experiment Manual or brochure

Book review Simulation exercises Portfolio

Letter of advice Performance Make a list

Newspaper article Presentation Experiment/hypothesis test

Lab report Debate Concept map

In-class writing exercise Interviews Assignments: Capstone course/project/experience

Annotated bibliography Fieldwork/internship/lab/clinical evaluation

Survey

Evaluate accuracy of … Testing Projects: group or individual

Research paper Written tests: objective Analyzing

Abstract Written tests: essay Case study

Internship/field experience/clinical report

Oral test Product analysis

Position paper Problem set Discussing

Critique Quizzes Discussion: classroom or on-line

Log Standardized assessment test of subject

Certification tests

Lab practical

Many assessment methods are applicable to more than one category

14

ELEMENTS OF ASSESSMENT AND PROGRAM REPORT

I. Assessment of Student Outcomes

Assessment is the process of gathering evidence of student learning, reviewing the evidence to

determine if students are learning what they are expected to learn, and using this evidence to

alter the direction of your course.

For example, you might ―map‖ certain questions on a test to specific learning objectives. After

administering a test, you would examine the students‘ performance on the test questions to

determine how well the students‘ are grasping the intended learning outcomes. If you determine

the performance is satisfactory, then you have evidence that the learning objective is being met.

If you determine the students‘ performance is below your expectations, you should use the

feedback to reevaluate the way the material is presented or review the concepts with students. It

is important to remember that the purpose of the assessment is to create a better teaching and

learning experience.

Students who know what is expected of them in terms of their learning have a framework for

learning and are more successful. Faculty who have a clear idea of what they want their students

to learn are able to align their instructional activities to these outcomes. In these two ways,

clearly articulated outcomes are essential to student learning. Outcomes assessment allows us to

systematically examine the alignment between student learning, instructional or institutional

expectations and instructional activities. To this end, we begin planning for outcomes assessment

with student learning outcomes. A student learning outcome (SLO) is defined as a specific,

measurable competency (knowledge, skills, values, or attitudes) that your students should be able

to demonstrate as a result of participation in a learning activity. SLOs reflect a shift from a focus

from ―What am I teaching‖ to ―What are my students learning?‖ SLOs can be expressed and

measured at the course, program or institutional level.

Course Assessments

Where do we start?

Every course should have a set of college-wide, common, core expectations for student learning.

These expectations are the most important things a student who passes the course should take

away from any section of the course. While individual instructors may add to this course, there

should be a shared understanding of the core skills and knowledge upon which the course is

based. It is these expectations which should be reflected on each course syllabus and which

should be used to determine student learning outcomes for the outcome assessment process.

Student learning outcomes are statements that specify what you want your students to know and

be able to do at the end of the course. For example, student learning outcomes can refer to

knowledge, practical skills, critical thinking skills, etc. that students are expected to develop or

learn.

15

What makes a good learning outcome?

A well-defined student learning outcome specifies actions by students that are observable,

measurable, and must be done by the students themselves. The crucial factor in determining if

your learning outcome is well-defined is whether or not the action taken by the students can be

measured. Do not focus on small details, but rather on general knowledge and/or skills you

expect your students to acquire through your course. Do not merely describe activities or lessons

from the course, but rather articulate the learning that will result from the course. Make sure your

statement is centered not on what you are going to teach them, but rather on what the student will

do. For example, ―upon completion of this course students will be able to identify all the critical

elections in 20th

Century America‖ as opposed to ―one objective of this course is to teach about

the critical elections in 20th

Century America.‖

Generally speaking, good learning outcomes are:

Learner centered

Key to the course‘s mission

Meaningful for faculty and students

Representative of a range of thinking skills

Measurable

First, and most importantly, good learning outcomes focus on what students can do instead of

the effort we put into teaching them. Second, college-wide outcomes must be essential to the

course’s mission, something that everyone teaching the course agrees is important. Avoid

outcomes that are idiosyncratic or tied to a particular instructor‘s approach to a course. Third,

design outcomes that are meaningful for faculty and students. If you cannot explain why a

certain outcome is important, it probably isn‘t very meaningful. Finally, outcomes often reflect a

range of thinking skills, from low level identification to higher level application of knowledge

or skills.

Good outcome are measurable in some way; they communicate what student learning will be

evaluated in the course. Often courses will have two levels of outcomes; some broader based

outcomes which reflect higher order thinking skills and broad topics, and some more narrow,

lower level thinking skills outcomes which are essential to reaching the broader outcomes.

If the course doesn‘t have expectations for student learning formulated as student learning

outcomes, the development of college-wide common core student learning outcomes maybe one

of the first outcomes of this process. The outcomes should become a standard part of the

syllabus.

When defining SLOs to assess, it is tempting to take the easy route and think only in terms of

learning outcomes that represent lower order skills because they will be simpler to evaluate.

Instead concentrate on the skills and knowledge which are essential for a student to be

considered competent at the end of the semester. While some lower order types of leaning

outcomes may be essential to reaching higher level outcomes, make sure that you define a range

of outcomes which reflect higher order, complex application tasks in addition to any essential

supporting learning outcomes which may reflect lower order thinking skills.

16

Lower order vs. higher order thinking skills

While basic recall of facts is important to any course, your assessment results will be more

meaningful if you have chosen a more complex skill. Moreover, it will likely reflect what is truly

important in your course. Often facts are important because we want students to be able to do

something with that information.

SLOs which reflect higher order thinking skills, use action verbs that are observable and

measurable, as well as ones that reflect higher order skills. Examples of such verbs are solve,

design, write, compare, apply, decide, draw, persuade, investigate, and evaluate.

Refer to the following possible outcomes for an information technology course:

Students will be able to correctly summarize the key differences between open and

closed source software development models.

Students will be able to evaluate the strengths and weaknesses of open and close source

software development models.

While the first outcome is certainly easier to achieve, the second one better represents what

students would have to do with the information in the real world. You will get more useful

information about student learning with the second SLO.

How do you write SLOs for a course or program?

A student learning outcome statement needs to specify who is to perform (student), what action

they are to take, and some result that must come from their action. A student learning outcomes

(SLOs) for a course/program should:

Be written in terms of what the student/graduate will be able to do at the end of the

course/program

Be limited to 2-5 outcomes

Keep them short and simple (KISS)

Make them specific, measurable, attainable, realistic, and timely (S.M.A.R.T)

Establish a target performance level for success (i.e. 70% will …)

Keep the assessment process manageable and meaningful (M&M)

Use Bloom‘s Taxonomy and active verbs (create, analyze, demonstrate, etc.)

Be written in the positive instead of the negative

Reflect measurable standards (benchmarks) or reflect the basic knowledge and skills that

the student will be held accountable for

Reflect a combination of higher order thinking skills and supporting or enabling skills

17

What are some basic examples of well-defined student learning outcomes?

Unclear student learning outcome statement:

The students will understand democracies.

The students will appreciate art from other cultures.

The students will learn about the law of relativity.

The above statements are not well-defined learning outcomes since they are not measurable.

However, these statements can be modified to become well-defined learning outcomes as

follows:

The students will be able to describe the major theories of democracy.

The students will be able to identify the characteristics of art from other cultures.

The students will be able to explain the major tenets of the law of relativity.

Sample Course Assessment Plan for SDV 108 College Survival Skills Goal/Objective Being

Assessed (SLO)

Evaluation method

(Expected Outcomes)

Findings Action to be Taken

Students will demonstrate

effective time management

skills 70% of the time.

Keep schedule in Daily

Planner

79.1% of students show

satisfactory or better planner

usage

Criterion has been

met, but important

dates needs to be

included in planner

Students will demonstrate-

appropriate comprehension of

course material 75% of the

time.

Weekly portfolio

assessment of reading,

writing & other

exercises

85.5% of students exhibit

satisfactory or better

apprehension of course material.

Criterion has been

met. Continue to

monitor

assignments and

exercises.

Students will utilize effective

study skills to successfully

complete course work 80% of

the time.

Class exercises &

research

essay/presentation

completion in portfolio

85% of students exhibit

satisfactory or better exercise

completion

Criterion has been

met. Continue to

monitor

assignments and

exercises.

Students will exhibit effective

note-taking skills 80% of the

time.

Class Note-Taking

exercises and evidence

of actual class notes in

portfolio

86.5% of student exhibit

satisfactory or better note-taking

skills

Criterion has been

met. Continue to

monitor note

Students will exhibit and

demonstrate personal

behavior that prepare them

for success with 80%

proficiency.

Class attendance &

participation, exhibition

of personal

responsibility using

rubric

84% of students regularly

attended class, participated in

class activities, turn in

assignment & bring required

materials to class

Criterion has been

met. Continue to

monitor attendance

and participation

Students will demonstrate

ability to use a computer and

information literacy skills

75% of the time.

Technology assignment

completion, library

resources, and email

sent to instructor

78% of students accessed library

databases, completed the library

orientation worksheet

assignment, and sent email to

instructor using blackboard

Criterion has been

met. Continue to

monitor Internet

technology

assignments and

library research

activities

18

General Education/Core Learning

Outcomes

English 111 Outcomes Upon successful completion of the course students

will be able to:

Communicate effectively orally and in writing

standard English

Formulate restricted, unified and precise

thesis statements

Organize essay content into introduction,

body, and conclusion paragraphs

Compose restricted, unified, and precise

topic sentences for paragraphs

Write unified and coherent paragraphs that

are well-developed with supporting

materials drawn from the literary text

Apply grammar and usage rules correctly

Choose appropriate diction

Write clear, precise sentences

Apply appropriate methods of mathematics to

solve problems

Comprehend and interpret reading materials

Explain basic literary terms in the genre of poetry,

fiction, and drama (for example, theme, imagery,

rhythm, figurative language, tone, character, plot,

etc.)

Understand and apply the methods, principles,

and concepts of the natural and social sciences

and the humanities

See above.

Understand the nature and value of the fine and

performing arts

English 112 requires this.

Use computer technology for communication and

information retrieval

Write research-based essays using secondary sources

to:

Synthesize several different sources into an

essay to support its thesis

Quote, summarize, and paraphrase

responsibly within that paper

Recognize and appreciate cultural diversity Students study the world‘s literature and write and

discuss a diversity of ideas.

19

Example of Action Plan

Action a department may take after

assessment

How specific courses planned to change

their courses after assessment

Change syllabi to prepare students for the

rigor of the course

English 150 Children‘s Literature professors

decided to emphasize the intellectual rigor and

copious reading in the class in the syllabus to

make students ―aware‖ that the assignments

and papers would be difficult.

Revise the course outcomes to include

more higher-order thinking, greater

intellectual rigor, and/or sufficiency

Many courses have merged similar outcomes,

omitted outcomes based on their lack of

intellectual rigor, and/or added language to

outcomes based on Bloom‘s Taxonomy of

high-order thinking.

Based on results from assessment, add or

reduce certain elements of the classroom

exercises

Using the equivalent of an item analysis, the

ELE 135 faculty noticed that many of the

questions answered incorrectly on their

assessment test were answered so because

students could not ―unlock meaning of

unknown words‖ based on prefixes and

suffixes. Hence, the faculty will investigate

how to emphasize word parts in ELE classes.

Obtain more consistency in large multi-

section courses

ITE 115 noticed that consistency in multi-

section courses is difficult, given that Franklin

Campus and Smithfield site do not have the

same resources. Although this analysis

delivers a negative truth, it also is

one worth noting.

Reduce grade inflation by linking test and

course grades to mastery of all outcomes

Assessment and analysis of Math 163 showed

that students‘ scores on the portion of the

exam that were common among all students

were not predictive of their final grade. This

portion, however, did not count toward the

final exam grade. Thus, it was speculated that

some students did not take that part of the

exam as seriously as the weighted part.

Increase contact with adjunct faculty Math 151 instructors also suggested that the

master syllabus may not communicate the

timing in which certain skills ought to be

taught and this would present problems,

especially to adjunct instructors who are not in

contact with faculty as much as full time

instructors.

Explore active learning strategies and

other teaching methods

In Physical Sciences 111, the instructor has:

Changed the sequence of course topics

20

for better flow

Introduced additional worksheets for

practice on skills

Spent more time discussing processes

De-emphasized memorization

Explore other ways of assessing

outcomes

The ENG 05 Developmental Reading/English

faculty decided that since they encourage

students to annotate their texts, the same

strategy ought to be applied when students are

being assessed. Because they were not aware

of this possibility, the faculty hypothesized,

students did not perform to their potential.

Explore technological enhancements

(labs, equipment, CD tutorial, etc.), using

the assessment evidence to support a

request for increased funding CIS

MKT 100 has discussed organizing and

cataloguing a library or videos relevant to the

course to better support visual learners.

Conduct a retreat or workshop for

instructors

Biology 101 examined their course and came

up with a plethora of questions. Based on this

analysis, the faculty desires to contact an

expert in assessment to find where and how to

proceed. The faculty emphasizes that their

desire to seek further help is linked to their

belief in assessment and its ability to enhance

student learning.

21

Program Assessments

Assessing student outcomes for programs is the most effective way to determine whether

PDCCC‘s programs are accomplishing the goals and objectives set forth in each program. A

careful analysis of the results of the students' assessment lets faculty and administration know

where improvements need to be made.

When doing annual student outcomes assessments, programs are asked to use multiple

assessment measures, of which at least one must be a direct measure. Faculty from each major is

asked to select any assessment methods that they believe will be effective in measuring whether

students achieved the goals of the program. The assessment of a major will give faculty vital

information concerning the program to be incorporated into the program review. Advisory

committees for OT programs are involved in reviewing curriculum. Some of the assessment

methods used are tests, competency checklists, rubrics, portfolio review, job placement rates,

employer surveys, oral examinations, written examinations, external certification examinations,

skills examinations, student surveys and panel reviews

In setting up student outcome assessments for any program or discipline, there are several

steps to follow:

1. State the program and general education goals/objectives.

2. Determine how and where each goal will be assessed using multiple methods (one

being a direct measure).

3. Record your results/findings.

4. Analyze the data collected by explain the results. Explain how your strategies to

improve your program worked or did not work. This may include a narrative

which describes the assessment process including how the data was analyzed and

the process in which assessment strategies were implemented.

5. Determine whether the goals of the program and general education goals are being

met, and state the actions taken (use past tense) or to be taken to address any concerns

or deficiencies. Close the loop.

22

Examples of Course, Program, and Administrative Unit Assessments

Direct Assessment Methods: Direct assessment methods give instructors measurable data to

study. Some examples are written exams, oral exams, performance assessments, standardized

tests, licensure exams, oral presentations, projects, demonstrations, case studies, simulations,

portfolios and, juried activities with outside panels.

Indirect Assessment Methods: Indirect assessment methods provide extra information that may

be used to make changes. Examples include questionnaires, interviews, focus groups, employer

satisfaction studies, observations of advisory boards, and job/transfer school placement data.

SAMPLE of the matrix used for Program Assessments

(Note: Analysis requires some explanation as to why the objective was met of not met.)

Goal/Objective being assessed Evaluation Methods Findings Actions taken or to be taken

Program Goal 1 Computer Analysis: Students will be able to diagnose, troubleshoot and repair computer system problems 75% of the time

Faculty observation during class and completion of task list test.

75% of the students received a satisfactory grade.

A handout on the requirements for a satisfactory grade was provided to students.

Analysis of Results (Explain how your strategies improved student success or did not improve student success):

This was an increase of 5% over last year (75% vs. 70%). This improvement appears to be due to the addition of simulation sessions on troubleshooting and repair. Program Goal 2 Students will demonstrate the ability to work effectively on a team 75% of the time.

Faculty observation during class-time devoted to group project work in (list course). Student written self-evaluation in (list course).

Faculty observed that 75% of the students demonstrated the ability to work effectively on a team. 90% of the self-evaluations indicated good understanding of effective teamwork

Have students critique a video showing a team at work and have them indicate which principles were well-employed and which were not.

Analysis of Results (Explain how your strategies improved student success or did not improve student success):

The group projects proficiency improved 8% over last year (75% vs. 67%). Part of this improvement appears to be due to using a rubric for group projects to identify strengths and weaknesses and early feedback to students so that the students know where they need to focus their attention to improve on their next group project. Program Goal 3 Students will demonstrate the ability to work effectively on a team 75% of the time.

Project assignments in ALL IST courses are completed in a timely manner. 80% of students will receive a favorable review of portfolios. The portfolios will be evaluated by program heads and the advisory committee. Where possible the portfolios will be available online.

All assignments have been completed in a timely manner per IST faculty. Portfolios were reviewed at the Spring IST meeting IST 226/129 – web sites IST179/180/216-review written procedures for troubleshooting, assembly, safety IST 202/CS200-engineering journals The Advisory Committee approved that all portfolios met IST standards.

No further action required at this time. Faculty continue to monitor and assure that assignments are completed in a timely manner.

Analysis of Results (Explain how your strategies improved student success or did not improve student success):

This year was an improvement (80% vs. 75%) from last year where students demonstrated the ability to work effectively in teams. Part of the improvement appears to be due to an increase emphasis and review by faculty on student portfolios in various program courses. This goal has met the benchmark goal of 75%.

23

Course Assessment Course Prefix and Number BIO 101 (Fictional)

Course Name: General Biology I

Instructor: All BIO 101 Instructors

Term and date: Fall 2010

Part I

(Completed sections of Part I should be submitted by August 27 to the Academic Dean for

review.)

List All Student Learning Outcomes Objectives from Course Syllabus (Note: You course objectives should come from your course syllabus)

Objective 1: Students will utilize the LRC for gathering information and completing four writing assignments to help them review current events. Objective 2: Students will be exposed to good laboratory techniques with regard to the proper care and use of laboratory equipment and supplies. Objective 3: Students will be able to identify cell types and label parts and know functions of each cell part. Objective 4: Students will be able to describe heredity components of living things. Objective 5: Students will be able to explain the process of protein synthesis. Objective 6: Students will be able to describe energy transformation as it applies to organisms. Objective 7: Students will be able to classify organisms. Objective 8: Students will be able to describe organic and inorganic chemical makeups of organisms.

List Any Primary Core Competencies Objectives Gen-Ed Objective 1: written communication (If you list any general education or core competencies here, you need to also show them below with some level of proficiency, evaluation method, and learning skill level or state which of the above objectives relate to the Gen-Ed Objective) Gen-Ed Objective 2: scientific reasoning

Note: At least 60-70% of student learning outcome objectives above should be assessed.

For each objective being assessed, list the

objective & measure of success (For objective

#3, students will be able to…with 70% of

proficiency)

Evaluation Method Learning Skill

(Bloom)

Obj. 1: Students will be able to complete 4 writing assignments with 70% of proficiency.

Read current event and write summary

Knowledge Comprehension Application Analysis

24

Synthesis Evaluation (Note: Do not list all skill levels for each objective, just the main one or two that you are evaluating.)

Obj. 2: Students will be able to identify lab safety rules and lab equipment with 100% proficiency. (Note: Do not set your objectives with 100% proficiency. If you do, you cannot improve.)

Write name of equipment from picture and write what’s done wrong in picture

Knowledge

Obj. 3: SWBAT label cells with 70% proficiency. Labeling parts of cell worksheet Knowledge

Obj. 4: SWBAT describe heredity and genetics (Note: Needs some level of proficiency. If you do not have some type of measurement, you cannot tell if you have met your objective.)

Create Punnett squares Synthesis

Obj. 5: SWBAT describe and explain protein synthesis with 70% proficiency.

Write steps of protein synthesis/draw and label each step and part

Comprehension

Obj. 6: SWBAT complete energy conversions and transformations with 70% proficiency.

Worksheet with pictures and energy transformation identification

Knowledge

Obj. 7: SWBAT put organisms into groups with 70% proficiency.

Test questions 4-6 on Test 1 Knowledge

Obj. 8: SWBAT list chemical symbols and the number of protons, neutrons and electrons for those essential to life with 70% proficiency.

Test questions 20-25 on Test 1 Knowledge

Actions Implemented: [What specific actions or new initiatives (if any) did you implement this year to improve your course? Why? ] Adopted a new textbook which focuses more on the course objectives and provides software that should be helpful to the students. (Note: You should be trying new things to improve your course)

General Education (Core Competencies):

(1) Oral Communication, (2) Written Communication, (3) Critical Thinking,(4) Scientific Reasoning,

(5) Quantitative Reasoning, and (6) Information Literacy

Sample Generic Grid for Mapping the Assessment (Make sure your grid shows a good balance of outcomes and enough attention to higher learning skills.)

Measure of success Evaluation Method Learning Skill (Bloom)

Objective 2: Students will …with

70% proficiency

Common questions 1-8 on Test 1 Knowledge

Objective 4: Students will … 80% of

the time

Common questions 20-35 on Test 2 Knowledge

Objective 5: Students will …with

75% proficiency

30-minute exam essay question,

scored by rubric

Comprehension, analysis

Objective 8: Students will …with

95% accuracy

Scored by rubric Comprehension, analysis, synthesis

Evaluation Method: (rubric, embedded test questions, project, lab test, journal, certification test,

portfolio)

Bloom’s Taxonomy:

(Note: higher level courses should have higher level thinking.)

Knowledge: Recall of previously learned facts

25

Words to use to assess recall: identify, define; describe, state, label, list, match, reproduce

Comprehension: Understanding what is meant

Words to use to assess comprehension: give examples of, classify, explain, describe,

summarize, outline, trace

Application: Use of previous knowledge to approach new situations or problems

Words to use to assess application: predict, construct, prepare, produce, show, use,

implement, design, show how

Analysis: Separate into component parts

Words to use to assess analysis: list the components parts of, break down, differentiate,

distinguish, diagram, illustrate, outline, subdivide, interpret, compare/contrast

Synthesis: Putting elements together so as to form a new concept

Words to use to assess synthesis: adapt, design, compare/contrast, categorize, compile,

assemble, rearrange, give evidence for, give reasons for, formulate, infer, generate,

integrate, plan

Evaluation: Judging by criteria

Words to use to assess evaluation: Develop criteria for, rank, prioritize, explain why you agree or

disagree, which is better, appraise, defend, judge, compare and contrast by criteria, review.

Complete at the end of the academic term

PART II (Completed sections of Part II should be submitted by January 7 to the Academic Dean. The Dean will

review and send to Director of Assessment & IR by January 7)

Findings/Results: (At the end of academic term, list each objective number that was assessed and results):

Objective __: Objective __: Objective __: Objective __: Objective __: Objective __: Objective __: Objective __:

Analysis & Evidence of Improvement: (To what factors for each objective did you attribute your findings/results to? Overall, what evidence of course improvement based on your analysis of results did you find?)

Objective __: Objective __: Objective __:

26

Objective __: Objective __: Objective __: Objective __: Objective __:

Action Taken to Modify Course to Improve Student Learning (based on results) and Why? (What will you do differently? Describe how the results obtained from the assessment will be used to improve student learning for objectives assessed. Why?) Objective __: Objective __: Objective __: Objective __: Objective __: Objective __: Objective __: Objective __:

Summary of Course Changes/Needs to Improve Student Learning: Changes made or needs for the course are the following: (1) Teaching methods (more homework, additional exercises, more emphasis on teamwork, providing review sessions, more hands-on, etc.) changes and/or changes to course syllabi: (2) Resources needed (equipment, software, student activities support for speakers, field trips, new textbook, tutors, etc.): (3) Policy change(s) (attendance, course pre-requisites, etc) that are needed to improve on student learning outcomes:

27

Course Assessment Course Prefix and Number :ENG 111 (Fictional)

Course Name: COLLEGE COMPOSITION I

Instructor: All ENG 111 Instructors

Term and date: FALL 2010

Part I

(Completed sections of Part I should be submitted by August 27 to the Academic Dean for

review.)

List All Student Learning Outcomes Objectives from Course Syllabus

Objective 1: The student will employ the writing process to compose various compositions: narrative,

expository, descriptive, and argumentative that are satisfactory on the collegiate-level in

focus, content, organization, style, and conventions.

Objective 2:

The student will read, interpret, analyze, and evaluate essays based on numerous human

experiences.

Objective 3:

The student will integrate their sources correctly in their research papers based on the revised

2009 Modern Language Association rules.

Objective 4:

The student will synthesize researched information to develop a well-documented research

paper.

Objective 5:

The student will define and apply the elements of logic and critical thinking to develop

argumentative writing.

Objective 6:

The student will use the computer to keyboard writings, exchange e-mails, complete

assignments.

Objective 7:

The student will evaluate his/her own writing and peers' using various writing strategies.

28

List Any Primary Core Competencies Objectives Gen-Ed Objective 1: The students will be able to employ written communication skills to develop compositions that are satisfactory in content, focus, organization, style and conventions 70% of the time. Gen-Ed Objective 2: The student will be able to demonstrate verbal and nonverbal effectiveness, appropriateness, and responsiveness 70% of the time.

Note: At least 60-70% of student learning outcome objectives above should be assessed.

For each objective being assessed, list the

objective & measure of success (For objective

#3, students will be able to…with 70% of

proficiency)

Evaluation Method Learning Skill

(Bloom)

For objective #1: Students will be able to compose a variety of compositions with 70% of proficiency in the areas of focus, content, organization, style, and conventions.

Instructor Generated Writing Rubric Knowledge Comprehension Application

For objective #2: Students will be able to read, interpret, analyze, and evaluate essays based on human experiences with 70% proficiency.

Instructor Generated Quizzes/Tests Comprehension Analysis Evaluation

For objective #3: Student will integrate their sources correctly in their research papers based on the revised 2009 Modern Language Association rules with 80% proficiency.

Instructor Generated MLA Rubric Knowledge Application

For objective #4: Students will be able to synthesize researched information to develop a well-documented research paper with 70% proficiency.

Instructor Generated Argumentative Research Paper Rubric

Synthesis

For objective #5: Students will be able to define and apply the elements of logic and critical thinking to develop argumentative writing with 80% proficiency.

Journal Responses - Argumentative Rubric

Knowledge Application

For objective #6: Students will be able to use the computer to keyboard writings, exchange e-mails, complete assignments 85% of the time.

Instructor Generated Writing and Computer Rubric

Knowledge Comprehension Application

For objective #7: Students will be able to evaluate his/her own writing and peers' using various writing strategies 85% of the time.

Peer and Self- Evaluation Writing Rubrics

Evaluation

Actions Implemented: [What specific actions or new initiatives (if any) did you implement this year to improve your course? Why?] I have implemented Service Learning into this course to allow students more opportunities to connect

29

real life to real learning, writing, reading, responses, and research. I have also embedded a collaborative teaching assignment that students become experts on the five characteristics of writing (focus, content, organization, style, and tone). In turn, they teach the concepts to the class based on their assigned learning styles groups.

PART II (Completed sections of Part II should be submitted by January 7 to the Academic Dean. The Dean will

review and send to Director of Assessment & IR by January 7)

Findings/Results: (List each objective number that was assessed and results):

Objective 1: Using a writing rubric, 88% of students employed the writing process to produce coherent, unified, and effectively developed compositions. Objective 2: Based on quizzes and tests, 85% of students were proficient on reading, interpreting, analyzing and evaluating essays.

Objective 3: Based on a MLA rubric developed by the instructor, 75% of students were proficient in integrating their sources correctly in their research papers using the revised 2009 Modern Language Association rules (MLA). Objective __: Objective 5: Using an argumentative rubric, 90% of students were able to compose an argumentative essay that addressed both sides and that was free of written expression and grammatical errors. Objective __: Objective __:

Analysis & Evidence of Improvement: (To what factors for each objective did you attribute your findings/results to? Overall, what evidence of course improvement based on your analysis of results did you find?)

Objective 1: Met Proficiency: The writing rubric showed that for Fall 2010, 88% of students were proficient in composing a variety of compositions. This was an increase of 4% over Fall 2009. This improvement appears to be due to more practice in-class writings that incorporate the writing process. Objective 2: Met Proficiency: Based on quizzes and tests, 85% of students were proficient on reading, interpreting, analyzing and evaluating essays. This compares to 80% in fall 2009. This increase in student success appears to be due to having weekly quizzes to the course. For the past few years this objective has been met or surpassed.

30

Objective 3: Below Proficiency: Based on a MLA rubric developed by the instructor, 75% of students were proficient in integrating their sources correctly in their research papers using the revised 2009 Modern Language Association rules (MLA). This was an improvement from last year which had a 65% proficiency. However, it is still below the proficiency level set by the department of 80% proficiency. The improvement appears to be from giving specific assignments that employ use of the revised 2009 MLA documentation style for each class assignment and allow students to practice with the new methods on a weekly basis. Objective __: Objective 5: Met Proficiency: Using an argumentative rubric, 95% of fall 2010 students were able to compose an argumentative essay that addressed both sides and that was free of written expression and grammatical errors. This compares to 90% for fall 2009. The improvement appears to be from adding in fall 2010 the strategy of having the instructor teach the necessity of reviewing and applying teacher comments, along with peer and self evaluation process. Objective __: Objective __:

Action Taken to Modify Course to Improve Student Learning (based on results) and Why? (What will you do differently? Describe how the results obtained from the assessment will be used to improve student learning for objectives assessed. Why?) Objective 1: No modifications are needed at this time. Objective 2: No modifications are needed at this time. Continuing doing weekly quizzes. Objective 3: A group power-point presentation and class discussion on use of MLA for research papers will be added to the course. Students are now receiving immediate feedback on their papers using the MLA rubric as a guide as to where a student weaknesses are. Objective __: Objective 5: No modifications are needed at this time. Objective __: Objective __:

Summary of Course Changes/Needs to Improve Student Learning: Changes made or needs for the course are the following: (1) Teaching methods (more homework, additional exercises, more emphasis on teamwork, providing review sessions, more hands-on, etc.) changes and/or changes to course syllabi:

31

Added weekly quizzes to increase student involvement and application (2) Resources needed (equipment, software, student activities support for speakers, field trips, new textbook, tutors, etc.): We adopted a new textbook in fall 2010 which has more chapters which focuses more on the course objectives. (3) Policy change(s) (attendance, course pre-requisites, etc) that are needed to improve on student learning outcomes: Completing ENG 03 was added as a pre-requisite to taking ENG 111.

32

Administrative of Justice (400) Assessment Plan (Fictional) Program Student Learning Outcomes Assessment Plan

(Note: Be sure to use multiple measures with at least one being a direct measure)

Objective Being Assessed

Evaluation method

Analysis Status (Met, Mixed,

Below Proficiency Level)

Objective 1: Graduating ADJ students will formulate an appreciation of ethical standards through their experiences (classroom/non-classroom) 70% of the time

Student proficiency of this learning outcome is formatively assessed throughout the program by using multiple measures such as midterm examinations, final examinations, and numerous Blackboard discussions. The primary measures are the final exam in ADJ 133, Ethics and the Criminal Justice Professional, and an assessment portfolio presented to the ADJ advisory committee and evaluated using a rubric. (Pilot May 2010 and 2011) Annual exit interview (starting April 2012)

2008-2009: Met Proficiency 2009-2010: Met Proficiency

Findings/Results:

Previous Year For summer 2009, proficiency was 94% based on a comprehensive exam vs. 92% for Summer 2008. For summer 2010, proficiency was 98% based on test scores and a partial portfolio review For 2009-10, the Advisory Committee evaluation of portfolios showed a 95% proficiency level. The rubric used showed no specific area of weaknesses.

2010-2011 (Enter the data results that you collected in the

current year)

Analysis & Evidence of Improvement:

[To what factors do you attribute your Findings/Results to? (For example, why did a particular objective improve from 57% in 2008 to 64% in 2009? Was this 7% increase due to a new initiative, pedagogy, new textbook, rubric, tools or activities, collaborations among program faculty or other discipline faculty?) Which factors will be discontinued, modified, or expanded? (For example, if the 7% increase was due in part to the use of a communication rubric designed in collaboration with ENG faculty … will AST faculty continue to use that rubric? If so, in which courses and how applied? Do AST faculty plan to continue collaboration next year to further refine the rubric? Etc…) Overall, what evidence of program improvement based on your analysis of results did you find?]

2008-2009: For 2008-09 proficiency improved from 92% in 2008-09 to 94% in 2009-2010. This improvement appears to have been enhanced by the increase use of Blackboard discussions on ethics. Using rubrics to assess student’s performance were also helpful to students for quick feedback and pointing out areas to improvement on weaknesses. The rubric showed no area across graduates that needed program modifications. In addition, students had to holistically consider professional ethical standards because of the comprehensive exam which was not required in 07/08. For 09/10 the addition of 35 ethical dilemmas will enhance the student’s ability to add a practical aspect to their learning. 2009-2010: Program improvement was evidenced by a partial portfolio review using a rubric that satisfied the Program Head and the Advisory Committee. The review showed a 95% proficiency level. This

33

improvement was attributed to the addition of the 35 dilemmas that added another practical dimension to the course and program. The advisory committee’s rubric used during the evaluation aided in pointing out any weaknesses that graduates had. The rubric showed no modifications needed at this time to the program.

Action Plan Modifications [What specific actions or new initiatives are you implementing to improve your education program? Why?] 2008-2009: Added 35 ethical dilemmas to the Ethics and the Criminal Justice Professional course to enhance the graduate’s ability to add a practical aspect to their learning. 2009-2010: Although the course and program goals were met (70%) for both years the use of the portfolio will continue to be presented to the ADJ advisory committee and evaluated using the course/portfolio rubric. (Pilot May 2010 and 2011) Annual exit interview (starting April 2012)

Objective Being

Assessed Evaluation Method Status

Objective 2: Communication: Graduating ADJ students will engage in listening skill improvement through their experiences (classroom/non-classroom) 70% of the time

The primary measure is in ADJ 227 and an assessment portfolio (case briefs and lecture summaries) presented to the ADJ advisory committee and evaluated using a rubric.(Pilot May 2010 and 2011 and annual exit interview starting April 2012) An ADJ Program Survey was also administered using a 3.0 as the proficiency benchmark.

2008-2009 Met Proficiency 2009-2010 Met Proficiency

Findings/Results:

Previous Year 2008-2009: For Spring 09, proficiency was 100% based on test scores and a partial portfolio review compared to 90% for the previous year. On an ADJ Program developed Survey, the 2009 ADJ graduates rated their proficiency as 3.2 on a 5-point scale with one being low. The benchmark for this objective was 3.0. 2009-2010: For spring 2010, proficiency was 88% based test scores and a partial portfolio review using a rubric by the Advisory committee. On an ADJ Program developed Survey, the 2010 ADJ graduates rated their proficiency as 3.5 on a 5-point scale with one being low. The benchmark for this objective was 3.0.

2010-2011

Analysis & Evidence of Improvement: 2008-2009:

34

The case briefs were effective and showed 100% proficiency vs. 90% in the previous year. The ADJ Survey also showed that 2009 ADJ graduates proficiency as 3.2 which was higher than the benchmark set of 3.0. This proved that students were listening. Students did not regard the lecture summaries as significant, however, to their test grade. For 2009-2010 the lecture summaries will have a course test grade value. 2009-2010: Lecture summaries were reviewed (88% proficiency vs. 100% for 2008-09). Still some students seemed to not care about the course grade value of the summaries. After a review of the makeup of the class non major and undecided students accounted for some of the decline in the total but not enough to fall below the 70%. The ADJ Survey showed an increase of proficiency for this objective (3.5 vs. 3.2 for 2009 ADJ graduates). These scores are above the benchmark set of 3.0.

Action Plan Modifications [What specific actions or new initiatives are you implementing to improve your education program? Why?] 2009-2010: Although the course and program goals were met (70%) for both years the use of the portfolio will continue to be presented to the Advisory Committee and evaluated using the course/portfolio rubric. Continue to monitor the non major and undecided students in the class (Pilot May 2010 and 2011 and annual exit interview starting April 2012)

Objective Being Assessed Evaluation Method Status Objective 3: Graduating ADJ students will formulate/improve supervisory skills through their experiences (classroom/non-classroom) 70% of the time

The primary measure is an assessment portfolio (of case studies and poster presentations) presented to the ADJ advisory committee and evaluated using a rubric. The primary examples will be obtained in ADJ 111, 112, and 231 for Police Science graduates and from ADJ 245 for Correction graduates (Pilot May 2010 and 2011 and an annual exit interview starting April 2012). An ADJ Program Survey was also administered using a 3.0 as the proficiency benchmark. Focus Groups are administered throughout the program.

2008-2009: Below Proficiency 2009-2010: Met proficiency

Findings/Results:

Previous Year 2008-2009: For Fall 2008, proficiency was not met 62% based on test scores and a partial review of portfolio material using the rubric On an ADJ Program Developed Survey, the 2010 ADJ graduates rated their level of proficiency as 2.5 on a 5-point scale with one being low. The benchmark for this objective was 3.0. The ADJ Focus group of graduates was conducted. It identified a need for more checking of portfolios by faculty to provide more feedback to graduates to keep them on track. 2009-2010: For Fall 2009, proficiency was met 81% based of test scores and a partial review of portfolio

2010-2011

35

material using the rubric On an ADJ Program Developed Survey, the 2010 ADJ graduates rated their level of proficiency as 3.2 on a 5-point scale with one being low. The benchmark for this objective was 3.0.

Analysis & Evidence of Improvement: 2008-2009: Based on the partial portfolio material students performed at a level that could be higher. Students performed only at a 62% level compared to the 70% benchmark set by the program head. On the ADJ Program Survey, the 2009 graduates scored 2.5 on this objective. This was below the benchmark of 3.0. The focus group showed a need for greater feedback as they go through the program. Because of the incompleteness/depth of the case studies the Program head will randomly check their portfolios during 2009-2010. Also for Fall 2009 the courses will require poster presentations to be completed. Overall program improvement should see an increase in improved supervisory skills. 2009-2010: The increase in competency test scores from 2008-2009 (62% proficiency) to 2009-2010 (81% proficiency) , the increase proficiency score on the ADJ Survey from 2.5 for 2009 graduates to 3.2 for 2010 graduates , and very positive comments about developing portfolios by the 2010 graduate focus group was the addition of poster presentations that gave students the in-depth analysis that is needed in the area of supervision. Poster presentations will in the future require all group members to participate in the presentation. Overall program improvement was documented by the depth of the material in the portfolio. The uninterrupted germane reading needs some more thought about how to measure and value this component. There were no interviews for spring 2010 since no graduates yet under this assessment plan.

Action Plan Modifications [What specific actions or new initiatives are you implementing to improve your education program? Why?] 2008-2009: Poster presentations are incorporated into ADJ courses listed above to promote supervisory skills. Students develop portfolios. Instructors are reviewing material as part of test scores and then the advisory panel/dean will review the material. 2009-2010: During the pilot phase of the assessment plan for the Administration of Justice Program the program head has been reading on cognitive loading theory and its relationship to memory. Since one of the main outcomes for Program students is thoughtfulness/deep thinking further readings on working, short and long term memory has convinced the program head to consider the following requirements for this course. Substitute some of the course projects with uninterrupted germane reading by Program students. This is based on the concept of memory leaching to sources of memory loading/storage devices and distractions experienced by Program students because of these devices. After a lengthy review of the literature on memory the program head has come to the realization that a majority of Program students may be losing the ability to think deeply because of the loading/storage devices and distractions. Because of the leaching of memory into these devices the brain is modifying itself because long term memory is not being used as much as in past, therefore a student may not rely on their long term memory

36

for the development of deep thoughts, for example, before these devices many people relied on their memory for telephone numbers and this exercised their memory just as reading has done for many years. In order for deep thought to occur the brain needs to have a great deal of information deposited to make the connections necessary to develop deep thought however these devices have made it more difficult to acquire/maintain information that has been externalized instead of being internalized.

… Objective Being Assessed Evaluation Method Status

Objective 9: Students will demonstrate an understanding of the concepts of normal and abnormal behavior, including focuses on the psychological and sociological aspects of criminal and other deviant behavior patterns with 80% proficiency

Student proficiency of this learning outcome is formatively assessed throughout the program by using multiple measures such as midterm examinations, and final examinations, The primary measure is the final exam in ADJ 247, Criminal Behavior. An ADJ Program Survey was also administered using a 3.0 as the proficiency benchmark.

2008-2009: Met Proficiency 2009-2010: Below Proficiency :

Findings/Results:

Previous Year For spring 2010 proficiency on a comprehensive test was 76% vs.92% for spring 2009 vs. 90% for spring 2008.

On an ADJ Program Developed Survey, the 2009 ADJ graduates rated their level of proficiency as 3.8 on a 5-point scale with one being low. The benchmark for this objective was 3.0.

On an ADJ Program Developed Survey, the 2010 ADJ graduates rated their level of proficiency as 2.9 on a 5-point scale with one being low. The benchmark for this objective was 3.0.

2010-2011

Analysis & Evidence of Improvement: 2008-2009: For this objective, student proficiency was 92% compared to 90% for 2007-08. This was supported by the ADJ Survey of 2009 graduates that rated their proficiency as 3.8. The benchmark for this objective was 3.0. Strategies used to improve performances from the previous year was an added class presentation and class paper on psychological and sociological aspects on criminal behavior in the ADJ 247 course. 2009-2010: Proficiency was only 76% for this year compared to 92% for 2008-09 and 90% for 2007-08. This drop was supported by the 2010 ADJ Survey on graduates which had a proficiency of 2.9 compared to 2009 ADJ graduates with 3.8. This is below the benchmark set of 3.0. This drop appears to be due to the adoption of a new textbook with a different publisher for ADJ 247 which does not address certain skills required for students in this program.

37

Action Plan Modifications [What specific actions or new initiatives are you implementing to improve your education program? Why?] 2008-2009: To improve public speaking, research skills, and knowledge about a specific aspect of criminal justice, a second presentation was added.

2009-2010: The program head has gone back to the original textbook, but with the new edition. Students will complete lecture summaries after each class then those summaries will be evaluated for areas of weakness and these areas will receive additional attention to be incorporated in the courses ADJ 107/247. Program Improvement Summary Report: [Based on the data analysis, how has your program improved this year? What are the strengths and weaknesses of your program? What new action plan initiatives are you planning for next year

(For example, at any time in the recent past did the AST faculty discuss using new office simulation software instead of the traditional use of textbook activities? Are faculty planning to implement something like that next year? Etc…)? What equipment/resource needs do you have for improving

your program?] 2009-2010:

After reviewing the findings/results for the year, we found that the Program has had some success

but could use some improvement in Objective 9. Students will complete lecture summaries after

each class then those summaries will be evaluated for areas of weakness and these areas will

receive additional attention to be incorporated in the courses ADJ 107/247. Time constraints will

not allow for the Advisory Committee to review the ‗lecture summaries‘ for this report; however

the review will take place as soon as the course is complete and the Advisory Committee can

meet. The Program Head has seen as a result of the ‗lecture summary‘ requirements and increase

in student questions, attention and recall. The program head has also approved going back to the

original publisher for the textbook used in ADJ 247.

Since this program subscribes to the philosophy of andragogy, the assessment strategy focuses

mainly on the affective domain which includes empirical research/observation to a great extent.

Based on the data, the Program will continue treating the Program students as adults, confer with

the Advisory Committee about the assessment test pilot, and make the necessary changes to the

Program with consequential testing.

38

Example: Annual Program Assessment__Administration of Justice (Fictional)

For Academic Year: 2010-2011

Program: Administration of Justice

Submission Date: May 15, 2011

MISSION:

Administration of Justice: The Administration of Justice program at ________ will provide

a foundation that combines a Christian values base with the theories, principles, and

practices necessary for a successful career in Administration of Justice.

OUTCOMES: Graduates will be able to:

1. Provide educational, remedial, and rehabilitative services to families and communities; 2. Organize communities and neighborhoods for social action; 3. Promote family and community interests in public, private, and governmental settings; 4. Maintain community resources of information, instruction, and assistance to all members of a

community; 5. Apply creative problem-solving on behalf of community members, especially those at risk of

violence in high crime areas; 6. Identify, analyze, and respond to problem situations involving civil rights, law enforcement, and

legal issues; and 7. Show competence in written and oral communications.

ASSESSMENTS:

Employer Internship Survey (Conducted in spring of Sr. internship)—Outcomes 1,2,3,&4

Student Internship Survey (Conducted in spring of senior internship)—Outcomes 1,2,3,& 4

Major Field Test (annually, senior capstone course)—Outcomes 5 & 6

Senior Project/Presentation/Portfolio (annually, senior capstone course)—Outcomes 1-7 LOCATION: All results of surveys, tests, and portfolio will be maintained in the office of the ADJ Program Head. DISSEMINATION/DISCUSSION:

The senior Project/Portfolio is evaluated according to a rubric by a team of evaluators including at least one department professor of Administration of Justice, one local police officer, and the director of security on campus.

Results of the surveys, major field tests, and portfolios are compiled, distributed, and discussed with the faculty and dean at department meetings.

Results are also distributed to the VP of Instruction & Student Development.

The ADJ Advisory Committee considers results at their fall annual meeting.

39

RESULTS: All ratings on the Employer Internship Survey met or exceeded last year’s results in the range of 90-96% for 2010-2011 as compared to 88-94% for 2009-2010 administration, except in one area. Business supervisors for the interns rated them lower than prior years (80%) in the area of communication regarding community resources. (Outcomes 1-4) Ratings on the Student Internship Survey were similar and met or exceeded the prior year, in the range of 95%=100% as compared to 94-99% for the prior year. However, students’ comments confirmed that some of them did not understand how to communicate the resources which are available in the communities of their internships. (Outcomes 1-4) Major Field Test scores were lower than the prior year (88% for 2010-2011; 92% for 2009-2010); however, they are within department’s established standard of success, set at 80% (Outcomes 5-6) Summary scores for the rubric scoring of the random sampling of Portfolios indicated that 80% of the students achieved at the acceptable or better levels for all of the performance criteria except in the area of “innovative education” for high-crime prone communities. This area continued to have a low score (50% acceptable or better) for the second year in a row. In addition, students’ oral presentations of projects indicated a less than acceptable (2.5 on a scale of 5 with one being low) overall score in oral communication for the second year in a row. MODIFICATIONS: For Outcomes #1-3—No modifications are recommended at this time. For Outcomes #4, 5, and 7__The department has met and revised the curriculum to include a field trip as a part of a capstone course (prior to their internship experiences) to a high crime area to collect information from a designated, high-need constituency in the community (seniors, disabled citizens, etc). Students will work as groups to design and communicate an “innovative education” plan regarding the safety resources available to the appointed constituency. TIMELINE, REQUIREMENTS, AND APPROVALS: The proposed capstone course curriculum and design (see attached outline) will go before the academic policies committee for approval in fall 2011. If approved by the committee, the faculty senate, the assistant VPAA and the VPAA, and if funding is provided for the field trip and training experience (see attached budget estimates) the additional course will be added for seniors in the fall 2011. Approval by Academic Dean __________________ Date: ____________________

40

Transfer AA&S Degree Programs (Business Administration, Education, Science, General Studies)

Assessment Plan WRITTEN

COMMUNICATION

INSTRUCTORS – 2008-2009: Mixed Proficiency

2009-2010: Mixed Proficiency

2010-2011: Given prior instruction, students will develop compositions that are satisfactory in content, focus, organization, style, and conventions 70% of the time.

Student proficiency of this learning outcome is formatively assessed throughout the program by using multiple measures such as midterm examinations, final examinations, numerous Blackboard discussion questions, as well as written papers across the program. The primary measures are the final exam in ENG 111 (College Composition I) using a rubric , the STAGE test developed by the VCCS. and PHI 115 (Capstone Course), as well as, a Writing Prompt and rubric developed by the Virginia Community College System (VCCS) using a six-point scale. Student proficiency of this outcome is formatively assessed in ENG 111 class by using the argumentative research essay, which will be evaluated using modified VCCS rubric.

Findings/Results Previous Years

For 2008-2009 proficiency, 76% of students enrolled in English 111 courses performed at a rate of 70% or better on their argumentative essays in areas of focus, content, organization, style, and conventions. For 2008-09, pending graduates scored 74.2% proficiency for written communication in the capstone course. For fall 2008, proficiency was 65% based on final exam test vs. 76% for fall 2007. For 2009-2010 proficiency, 84% of students enrolled in English 111 courses performed at a rate of 70% or better on their argumentative essays in areas of focus, content, organization, style, and conventions.

For fall 2009, proficiency was 64% based on final exam test. For 2008 graduates, the weakest area on the writing prompt using the VCCS rubric was in Organization (4.55) and the highest in conventions (5.35).

2010-2011 (Add your new data for 2010-2011)

2010 graduates had a mean score of 6.89 (ten-point scale with 1 being low) on the communication section of the STAGE test. Using item analysis, the weakest area was demonstrating the ability to use standard English.

2010 graduates were 85% proficient (based on 80% accuracy) in the capstone course module on written communication. Using a rubric, the weakest area (30% proficiency) was in Organization: Organize content with effective transitions and effective beginning and ending paragraphs.

2010 graduates using a rubric developed by English faculty and the VCCS had mean scores (from 1-6 with 1 being low) of 4.28 on rhetorical knowledge, 4.06 on critical thinking, 4.10 on organization, 4.08 on content & clarity, 4.10 on style, and 3.88 on conventions. The weakest area (3.88) is on conventions.

Analysis & Evidence of Improvement:

[To what factors do you attribute your Findings/Results to? (For example, why did a particular objective improve from 57% in 2008 to 64% in 2009? Was this 7% increase due to a new initiative, pedagogy, new textbook, rubric, tools or activities, collaborations among program faculty or other discipline faculty?) Which factors will be discontinued, modified, or expanded? (For example, if the 7% increase was due in part to the use of a communication rubric designed in collaboration with ENG faculty … will AST faculty continue to use that rubric? If so, in which courses and how applied? Do AST faculty plan to continue collaboration next year to further refine the rubric? Etc…) Overall, what evidence of program improvement based on your analysis of results did you find?]

41

2008-2009: Students have a mastery of focus, style, and content development thus providing more collegiate-level writing throughout the program. Proficiency was 76% on the argumentative essays using a rubric. This was supported by the 74% proficiency on the writing module in the capstone course. Both of these measures meet the proficiency level of 70% set by the department. However, on comprehensive tests, student proficiency was only 64% in fall 2008. This competency is being expanded by having students' writing in non-English courses in the area of focus, content, organization, style, and conventions, Non-English faculty ( history, religion, technology, and psychology) incorporated writing into their disciplines and generate a method of evaluation such as a rubric to provide students feedback on their writing in addition to the content in the areas of focus, organization, style, conventions, along with content. The English faculty will revise the departmental rubric that was used to evaluate final exams and the argumentative research essay so that common rubrics are used amongst all English faculty for each type of essay assigned. 2009-2010: Students have a mastery of focus, content, style, and organization thus providing more collegiate-level writing throughout the program. The 8% proficiency improvement for students enrolled in English 111 on their argumentative essays in areas of focus, content, organization, style, and conventions (84% proficiency in 2009-10 vs. 76% in 2008-09) and the increase in performance in the capstone course (85% in 2009-10 vs. 74% proficiency in 2008-09) was a result of inclusion of collaborative argumentative writing exercises embedded into English 111 courses, adoption of new textbooks for English 111, employment of journal writing into English 111, incorporation of peer and instructor evaluations, and learning resource presentations which offered resources where students could obtain information to improve writing. Students’ proficiency on tests still appears to be an issue with only 65% proficiency in fall 2009. This is, however, a 1% improvement over last year. To improve areas of organization as noted in 2008 graduates' writing, English instructors used practice exercises to teach students how to embed transitional words into the content of their writing, how to analyze writing with and without transitions to examine the effectiveness of using transitions to aid organizational skills, and how to develop topic sentences that help structure the essay. In addition, demonstrations of how to use graphic organizers as means to help develop ideas in an organized manner were also used along with new textbook adoptions for English 111 which provided exemplars of each genre of writing by which students used to enhance their organization. To improve the use of Standard American English, Modern Language Association formatting of document was required for final submissions of work in all English courses and the writing handbook and newly adopted text with exemplars were used for English 111 to help students improve word choice and use of Standard American English. The 11% percent increase in graduates' proficiency in the capstone course module (85% in 2009-10 vs. 74% in 2008-09) on written communication was a result of requiring students in English 111 to use argumentation and critical thinking skills to help develop writing. In addition, in 2008, English 03 - Preparing for College Writing II was added as a bridge course between English 01 and English 111 to help teach more complex foundational

42

writing and grammar skills to students so that they would be better equipped to handle the rigor of English 111. 2010-2011:

Action Plan Modifications [What specific actions or new initiatives are you implementing to improve your education program? Why?] 2008-2009: The English faculty added English 03, a bridge developmental course between English 01 and English 111, to provide students with more complex skills in writing, grammar, and research.

The English faculty and psychology faculty conferred with students about their writing and areas of

needed improvement at least one time a semester.

In all courses at PDC, faculty have the option to have a Library Help tab added to their Blackboard

course which contains specific materials and resources that aid students in developing college-level

writing for various disciplines. 2009-2010: The English faculty adopted new textbooks of which contains exemplars of each genre of writing along with essays readings to strengthen each student's focus, content, organization, style, and conventions. The English faculty added journals as a means of reflection, critical thinking, and writing to provide more opportunities for students to practice their writings in English 111.

English faculty used graphic organizers and other visuals to improve organizational skills of writers.

English faculty obtained more external resources such as books, multi-media resources, and other

external resources to aid students in development strong argumentative writing skills.

English faculty conferred with each student at least once a semester about how to improve his/her

collegiate-level writing to level of proficiency. In the capstone course, redesigned assignments to demonstrate clearly the correlation of organization and transitioning of concepts. 2010-2011:

43

Example Administrative Unit: Assessment and IR (Use multiple measures direct and indirect)

(Note: This example does not show action plans for any objective not meeting benchmark) Goal/Objective Assessment

Measurements Results Evidence of Improvement & Analysis

Research and Assessment is adequate & appropriate to the college mission with a benchmark performance rating of 3.0 or higher

Faculty & Staff Survey based on a 5-pt. scale from 1-5 with 1 being low GAP Analysis Focus Groups

For 2006-07 the performance rating was 3.7 compared to 4.0 for 2010-11; A GAP Analysis on the college’s policies and procedures was done in 2006 which identified a need to have better ways to track student groups for course completion rates, graduation rates, grade distribution Two Focus Groups composed of (1) support services staff and (2) classified staff was done in 2006-07 to identify college weaknesses. There were none reported for assessment and research.

(Analysis Status: Met, Mixed, Below Benchmark) Met Benchmark This goal has improved when comparing 2006-07 survey data with 2010-11 data by +0.3 points. Strategies that contributed to this improvement included: (1)The Office of Assessment & IR wrote an Achieving the Dream grant for over $450,000 to help the college in its strategies to improve on student success based on data and assessment support; (2) The Assessment & IR office has updated its SAS statistical software Package from 8.2 to 9.0; (3) The Office of Assessment has a color printer to better present data to its various stakeholders; (4) Various queries and crystal reports were developed in PeopleSoft to provide the data needed; and (5) the VCCS and SCHEV worked together to make data more available to colleges in regard to transfer stakeholders.

Action Plan/Modifications: No modifications are recommended at this time.

Assessment results are used for the improvement of programs and services with a benchmark performance rating of 3.0 or higher

Faculty & Staff Survey based on a 5-pt. scale from 1-5 with 1 being low

For 2006-07 the performance rating was 3.2 compared to 3.7 for 2010-11;

Met Benchmark This goal has improved when comparing 2006-07 survey data with 2010-11 data by +0.5 points. Strategies used to make the improvements included: (1) placing program/services data on the college web page, (2) placing program working documents on a common drive for faculty to access and to enter data, (3) professional development training provided to faculty on assessment, as well as face-to-face mentoring, and (4) had program faculty do annual program

44

assessments instead of waiting for the 5-year Program Assessment Report.

Action Plan/Modifications: No modifications are recommended at this time.

Relevant information from IR, such as that required for decision-making, is readily available with a benchmark performance rating of 3.0 or higher

Faculty & Staff Survey based on a 5-pt. scale from 1-5 with 1 being low

For 2006-07 the performance rating was 3.4 compared to 3.7 for 2010-11;

Met Benchmark This goal has improved when comparing 2006-07 survey data with 2010-11 data by +0.3 points. Strategies used to make the improvements included: (1) developed an Assessment & IR web page (2) placing the Fact Book on the college Assessment & IR web site, (3) placing the Benchmark Report on the college Assessment & IR web site, (4) used Achieving the Dream data in making changes such as adding another level in Developmental English and in increasing the placement test cut-scores, and (5) placed a number of resource documents and assessment tools on the Assessment & IR web site to assist the administration in making decisions.

Action Plan/Modifications: No modifications are recommended at this time.

Any request made to Assessment and IR were completed in a timely manner with a benchmark performance rating of 3.0 or higher

Faculty & Staff Survey based on a 5-pt. scale from 1-5 with 1 being low; Percent of IR reports completed by requested time Zero Complaints to supervisor or president

For 2006-07 the performance rating was 3.6 compared to 3.8 for 2010-11; For 2007, 92% (22/24) of reports were completed by the requested data compared to 97% (28/29) for 2009; For 2009-10, there were no complaints made to supervisor or president for not completing data requests in a timely manner

Met Benchmark This goal has improved when comparing 2006-07 survey data (3.6 rating) with 2010-11 data (3.8 rating) by +0.2 points. The percentage of assessment and IR reports has also improved from 95% (22/24) being completed by the requested time in 2007 to 97% (28/29) in 2009. There were also no complaints for 2009-10.

Action Plan/Modifications: No modifications are recommended at this time.

Assessment and IR plays a significant role in making college

Faculty & Staff Survey based on a 5-pt. scale from 1-

For 2006-07 the performance rating was 3.6 compared to 3.8 for 2010-11;

Met Benchmark This goal has improved when comparing

45

improvements with a benchmark performance rating of 3.0 or higher

5 with 1 being low 2006-07 survey data with 2010-11 data by +0.2 points.

Action Plan/Modifications: No modifications are recommended at this time.

Example Administrative Unit: Admissions (Fictional)

Goal/Objective Assessment Measurements

Multiple Assessments

Results Evidence of Improvement & Analysis

The College’s admission policies seeking are efficient for students and staff. A benchmark of 3.0 was established.

Student Survey based on a 5-pt. scale from 1-5 with 1 being low Focus Groups Gap Analysis

Although achieving an overall rating of 3.2 surveyed students indicated the College’s Admissions process is cumbersome and should be streamlined. Community members were invited to participate in an Admissions focus group. Findings indicate that most community members are unaware of the steps required to enter the College. When apprised of the steps involved participants indicated that the process should be streamlined. Admissions staff reviewed all policies and procedures relating to College admission. Possible efficiencies could be achieved by better use of available technologies.

(Analysis Status: Met, Mixed, Below Benchmark) Mixed Benchmark The achieved survey scores exceeded the benchmark. Strategies that lead these results: 1) Customer service training provided to all Admissions Office staff. 2) A computer workstation was added to the Admissions Office for students to input application data. 3) A brochure detailing the admissions process was prepared and disseminated locally. Although the overall benchmark of 3.0 was exceeded it was indicated that the College’s admission process could be improved and streamlined.

Action Plan/Modification (If status is mixed or below benchmark, some type of actions should occur.)

46

Assessment Techniques

There are many techniques that may be used to assess student learning outcomes. In a number of

cases, these assessment techniques may be embedded in course assignments or activities as

measures of students‘ achievement of program goals as well as their attainment of the college's

general education goals.

Capstone Courses Capstone courses are designed to enable students to review, evaluate, integrate, and synthesize

information and skills gained from other courses in the program or major. These courses are the

optimum place to assess many program or major goals and general education goals. A capstone

course is one which completing students take as a culminating experience that gives them the

opportunity to tie together the knowledge and skills of other program courses. If your program

has such a course, you may want to consider the performance in this type of capstone course as

an assessment method. Likewise, some programs assign a capstone project which can be

evaluated.

Criteria-Referenced Statement of Summative Learning These statements show that graduates are learning a skill that‘s important in their disciplines. For

example a statement might note that, ―89 percent of our AAS degree recipients in Allied

Health/Medical Tech solved 20 simulated tasks concerning drug side-effects using the

Physician’s Desk Reference.‖

Internships, Field experiences, Clinical Evaluations Internships, field, or clinical experiences are also ideal for assessing many program or major and

general education goals. When these occur at the end of the program or major, they often serve

as capstone experiences. It is especially useful to have external experts assess the performance

of your students.

Authentic Assessment In some courses, opportunities can be found to ask students to engage in a simulation of a real-

life problem that they must solve using the knowledge and skills they have gained in the course.

A single project can be structured to assess both mastery of course content and attainment of

program or major goals as well as certain general education goals such as communication skills,

life-long learning skills, critical thinking skills, and social and education values. For example,

students might be asked to assume the role of a city council member who must make a decision

concerning a controversial issue. Students might then be asked to research both sides of the

issue and to deliver a persuasive speech or to write an action plan.

Ill-defined or Ill-structured problems

An ill-defined problem is one that is not highly structured and cannot be resolved with a high

degree of certainty. Experts may disagree about the best solution. Examples: determining what

really happened at Waco or solving the nuclear waste storage problem or predicting the effect of

global warming or deciding if there is such a thing as global warming. Dealing with ill-defined

problems requires the integration of many skills, abilities, areas of knowledge.

47

Portfolios An accumulation of student-produced work, a portfolio may be designed to assess a student‘s

attainment of program or major goals. The same portfolio may also be used to assess general

education goals such as communication skills or the development of skills to enhance life-long

learning, such as the ability to use the library and other appropriate sources to retrieve

information. Portfolios that contain early or unrevised work as well as later or revised work can

assess the growth of skill development. Rubrics to judge portfolios must be clear and shared

with the student.

The Advisory Committee (who are working professionals in the field) judged the work in the

portfolios using detailed criteria. This process assessed the individual student‘s work so that the

student could remedy any problem areas during the last semester, and the analysis of the

portfolios as a group indicated areas of concern for the program The students then had

something tangible to take with them on job interviews which showcased their work.

Curriculum Analysis Review

This is a common assessment activity used by a number of occupational/technical programs.

The Advisory Committee is particularly useful in curriculum review because they are generally

practicing in the field and are aware of advances or changes. Often the advisory committee can

give valuable insight by reviewing the goals and objectives to help plan future directions of a

program. Tying a curriculum to a national standard may be a particularly valuable assessment

technique.

The advantage of using this as one aspect of a program‘s assessment is that by using the advisory

committee, local business/industry are getting a voice in whether the curriculum is meeting their

needs. It is also an inexpensive assessment tool.

However, keep in mind that although we need to be sensitive to the needs of local business and

industry, it is our obligation to prepare students to work outside our service area as well as within

our own region. Generally, we can assume that the skills and knowledge needed in a certain

field in our own region will serve a student well anywhere, but there may be instances where that

does not prove to be the case.

Grades

Grades can be used to assess student learning by using primary trait analysis (PTA) to identify

the factors that count for scoring and explicitly stating the criteria for the evaluation of the

assignment, project, presentation, product in the form of a rubric.

Course-embedded Assessment

Program or major goals and general education goals may be assessed through assignments

embedded in required courses. For example, writing assignments, such as summaries or reports,

and oral presentations may be used to assess student' mastery of course content as well as their

writing, reading, critical thinking or speaking skills and use of the library or other information

source. With some planning, a single assignment or project can be designed to assess a number

48

of different program or major goals as well as general education goals.

Critical Incidents

Students can be asked to describe an incident, either real or imagined, that illustrates or

illuminates key concepts or principles. An explanation of the concepts or principles illustrated

should accompany the description of the incident.

Case Study Presented with a realistic example of an application in the field, students must

respond with an analysis that demonstrates their mastery of course content and their ability to

apply the information and skills they have learned. A case study is an examination of a specific

phenomenon such as a program, an event, a person, a process, an institution, or a social group.

The end product of a case study is a rich, thick description of the phenomenon being studied that

illuminates the student's understanding of the phenomenon through the application of the

knowledge and skills they have gained.

Journals Journals or learning logs have been used in composition courses for years as a tool for increasing

student writing and motivation for writing and for assessing students' writing skills. However, a

journal that focuses on students‘ social and educational attitudes and values may be also useful to

assess students‘ achievement of general education goals. Journals may also be used to assess

student attainment of program or major goals.

Writing Samples Writing assignments can be used as a measure of students' mastery of course content

and attainment of program or major goals. Such assignments may also be used as a direct

measure of the general education communication skills goal as well as an indirect assessment

of critical thinking skills. Examples of writing samples include essays, research or term papers,

answers to essay questions on tests, book reports, summaries, lab reports, and the like.

Oral Presentations/ Oral Exams Depending on the nature and content of the course, oral presentations can be tailored not only to

assess students' mastery of course content but also their attainment of general education goals

such as critical thinking, general knowledge and historical consciousness, understanding the

impact of science and technology, and educational and social values. Oral presentations based

on course content can be used as a direct measure of students‘ communication skills.

Certification Tests

Programs in which a student must pass a certification examination in order to be certified to

work in the field, such as nursing, may want to consider using the results of that test as an

assessment technique.

One advantage of doing that is that successful results demonstrate credibility of the curriculum.

One disadvantage is that many organizations will not disclose students‘ results to the college

(although individual students might).

49

Exit Interviews

There are different types of exit interviews, but they commonly fall into two categories. In one

type of exit interview the program head and students discuss topics similar to those found on

student surveys. Topics can be very detailed and may result in information that you hadn‘t

thought to request. Sometimes students will say things that they do not wish to put in writing.

The other type of exit interview is actually more like an oral examination. (Call it an exit

interview has the advantage of not scaring students to death.) This method has been used very

successfully by the Administration of Justice program, where the interviews are conducted by a

panel made up of advisory committee members. It has the advantage of giving students practice

in the kind of interviews that they face for the hiring process and future promotion boards and

also assessed their proficiency in both oral communication and knowledge of their subject area.

Focus Groups

Focus groups are structured but informal discussions with small groups of students. Students

may be asked about issues that are pertinent to the program. Focus groups can also be conducted

with faculty, advisory committees, administrators and other employees.

External Evaluation/Review

This is a type of peer review where a consultant(s) from either business or another institution

examines a program from an outside perspective. This may involve such things as visiting

classes, interviewing faculty and students, interviewing advisory committee members, examining

curriculum goals and objectives, reviewing final exams, and interviewing local business and

industry. This method provides the opportunity for the exchange of ideas with a faculty member

of another institution.

Course Tests and Exams Common test questions drawn from course content and included on tests and exams in all

sections of the course can be used to assess both program or major goals and some general

education objectives. A locally developed test gives you the opportunity to determine if specific

desired outcomes are being successfully attained. It can be tailored to meet the objectives of

your program. However, preparing a test takes a great deal of preparation and study.

Rubrics

For scoring consistency with longer open-ended assignments such as essays, research papers, or

performances, a rubric should be developed. A rubric is a criterion based scoring tool that

specifies levels of achievement (e.g. exemplary, satisfactory, and unsatisfactory) for each

dimension of the outcome. As part of the rubric, criteria are provided that describe what

constitutes the different levels of achievement.

There are two major types of rubrics: holistic and dimensional (analytic) also known as primary

trait rubric. Both detail the particular qualities that separate excellent from poor student work

along a spectrum, but the first groups the dimensions together, while the second keeps them

separate.

The holistic rubric looks at the instrument as a whole; students receive one overall score based

50

on a pre-dimension scheme used by everyone. The dimensional (analytic) rubric yields sub-

scores for each dimension, as well as a cumulative score which is the sum, either weighted or un-

weighted, of the dimensional scores.

Each type of rubric has its strengths and weaknesses. Holistic rubrics allow you to look at a

student‘s overall performance, and often it corresponds better to the grade that pops into our

heads immediately after we finish looking at the student work. The dimensional (analytic) rubric

provides more information about what is working and what is not. For example, perhaps students

are doing a good job with learning the mechanics or writing, but not so well with learning

writing development. A dimensional rubric will provide information with this level of detail,

whereas a holistic rubric will not.

Regardless of the type of rubric, it is important tht it be shared with students well before the

assessment is administered. It is unreasonable to expect students to perform well on an

assessment if they do not have a clear understanding of the standards being used to evaluate it.

Surveys Surveys may be used to assess the degree to which students perceived that they have attained

program or major goals as well as certain general education goals. Items that elicit this

information may be included on surveys developed by program or major faculty and

administered to current and/or prior students and on surveys sent to employers of program or

major graduates.

The use of surveys is a way to gain information that may directly impact a program. There are

many types of surveys. The ones most often used are graduate surveys, employer surveys and

student surveys. Surveys allow you to get direct feedback from a number of perspectives such as

employers and graduates. Results sometimes raise issues that would not be apparent in other

types of assessment.

One disadvantage is that it is often time-consuming and expensive. It requires careful planning

since a survey that is not thought through thoroughly may give you little useful information.

Standardized Tests

Standardized tests are nationally normed and may also be used to assess students‘ perception of

their attainment of general education goals. These tests best assess reading comprehension,

critical thinking, scientific reasoning, the ability to solve math problems, and writing skills such

as knowledge of grammar and correct usage. Additionally, there are major field tests which may

be used to assess student learning.

When administered pre and post, standardized tests can be an effective way to measure

achievement in a particular area. They have the advantage of credibility since they are nationally

normed. However, these tests are often expensive and do not always match well with the

curriculum. Our use of standardized tests in assessment has been limited in the past. We have

found that although it is good for detecting general problem areas, it is sometimes quite difficult

to discern more specific areas needing attention.

51

How to Design Rubrics for Scoring Essays, Projects, and Performances

Follow These Steps

1. Decide whether you want a holistic or analytic

rubric.

2. Construct a primary trait scale (a rubric).

3. Obtain consistency in instructions and conditions.

4. Norm the scorers.

A scoring rubric applied consistently by faculty teaching the course is a good way to assess

essays, projects, and performances. A rubric describes the primary traits of a high-level essay or

project, a poor essay or project, and the levels in between. That is, a rubric lists the criteria for an

A, a B, a C, etc., or for a score of 6, 5, 4, etc.—depending on how many levels of differentiation

are desired. Instructors use the rubric to score the essay, project, or performance.

1. Decide whether you want a holistic or analytic rubric.

An analytic rubric measures each part of the student work separately; a holistic rubric combines

them. To illustrate, here are analytic and holistic rubrics to assess Spanish journals in a beginning

Spanish course

Analytic Rubric for Spanish Journal Holistic Rubric for Spanish Journal

Comprehensibility

4. Entries are completely understandable.

3. Entries are usually understandable.

2. Entries are difficult to understand.

1. Majority of entries are incomprehensible.

Usage

4. Although there a few errors, verb tenses, sentence

structure, and vocabulary are correctly used.

3. Some use of appropriate verb tenses and correct

sentence structure and vocabulary, but incorrect

usage or vocabulary interfere.

2. Many errors make comprehension difficult.

1. The majority of entries are incomprehensible.

Risk Taking

4. Student has taken some chances, employing

sentence

structures on the edge of what we have been

studying.

3. Student writes mostly safe entries, but is

generally

current with the textbook.

2. Student writes only safe entries, and is not current

with the textbook.

1. Student writes only simple structures.

Note that several traits (comprehensibility, usage,

risk taking, and variety of subject and form) have

been combined into a single scale.

4. The content of the journal is comprehensible.

Although there are errors, verb tenses, sentence

structure, and vocabulary are correctly used. The

author has taken some chances, employing sentence

structures or expressing thoughts that are on the

edge of what we have been studying. The entries are

varied in subject and form.

3. There is some use of appropriate verb tenses and

correct Spanish sentence structure and vocabulary,

but incorrect usage or vocabulary interferes with the

reader‘s comprehension.

2. The reader finds many of the entries difficult to

understand, or many entries are simplistic or

repetitious.

1. The majority of entries are incomprehensible.

Source of holistic rubric: Barbara Walvoord and

Virginia Anderson, Effective Grading: A Tool for

Learning and Assessment, 1998.

52

Variety

4. Entries are highly varied in subject and form.

3. Entries are somewhat varied in subject and form.

2. Entries show only a little variety in subject and

form.

1. Entries show no variety in subject and form.

EXAMPLE: MTH 163 Proficiency Learning Objectives Rubric

Using Comprehensive Test

Sample size consisted of ______ students.

Number of

students

below

Proficiency

1 point

Number of

students

at

Proficiency

2 points

Number of

students

above

Proficiency

3 points

1. Student’s ability to use mathematical logic and reasoning to solve content related problems

2. Student’s ability to interpret and use content related formulas

3. Student’s ability to make inferences based on interpretation of graphs, tables, and/or schematics

4. Student’s ability to solve content related problems by using algebra, geometry, and/or statistics

5. Student’s ability to determine reasonableness based on estimated answers

6. Student’s ability to recognize and communicate appropriate methods to solve content related problems

7. Student’s ability to represent mathematical information numerically, symbolically, and/or visually with graphs and charts.

Above Proficiency indicates that a student correctly answered 90% or more of Quantitative Reasoning Objectives items.

At Proficiency indicates that a student correctly answered between 70% and 90% of Quantitative Reasoning Objectives items.

Below Proficiency indicates that a student correctly answered 70% or less of Quantitative Reasoning Objectives items.

53

Public Speaking Assessment Rubric Verbal Effectiveness – 50 points

Idea development, use of language, and the organization of ideas are effectively used to achieve a purpose.

Advanced Developing Emerging

A. Ideas are clearly organized,

developed, and supported to

achieve a purpose; the purpose

is clear.

B. The introduction gets the

attention of the audience.

C. Main points are clear and

organized effectively.

D. Supporting material is original,

logical, and relevant.

E. Smooth transitions are used.

F. The conclusion is satisfying.

G. Language choices are vivid and

precise.

H. Material is developed for an oral

rather than a written

presentation.

A. The Main Idea is evident, but

the organizational structure

may need to be strengthened;

ideas may not always flow

smoothly.

B. The introduction may not be

well-developed.

C. Main points are not always

clear.

D. Supporting material may lack

in originality or adequate

development.

E. Transitions may be awkward.

F. The conclusion may need

additional development.

G. Language is appropriate, but

word choices are not

particularly vivid or precise.

A. Idea ―seeds‖ have not yet

germinated; ideas may not be

focused or developed; the main

purpose is not clear.

B. The introduction is

underdeveloped or irrelevant.

C. Main points are difficult to

identify.

D. Inaccurate, generalized, or

inappropriate supporting

material may be used.

E. Transitions may be needed.

F. The conclusion is abrupt or

limited.

G. Language choices may be

limited, peppered with slang or

jargon, too complex, or too

dull.

Nonverbal Effectiveness – 50 points

The nonverbal message supports and is consistent with the verbal message.

Advanced Developing Emerging

A. The delivery is natural,

confident, and enhances the

message – posture, eye contact,

smooth gestures, facial

expressions, volume, pace, etc.

indicate confidence, a

commitment to the topic, and a

willingness to communicate.

B. The vocal tone, delivery style,

and clothing are consistent with

the message.

C. Limited filler words (―ums‖) are

used.

D. Clear articulation and

pronunciation are used.

A. The delivery generally seems

effective – however, effective

use of volume, eye contact,

vocal control, etc. may not be

consistent; some hesitancy may

be observed.

B. Vocal tone, facial expressions,

clothing, and other nonverbal

expressions do not detract

significantly from the message.

C. Filler words are not distracting.

D. Generally, articulation and

pronunciation are clear.

E. Over dependence on notes may

be observed.

A. The delivery detracts from the

message; eye contact may be

very limited; the presenter may

tend to look at the floor,

mumble, speak inaudibly,

fidget, or read most or all of

the speech; gestures and

movements may be jerky or

excessive.

B. The delivery may appear

inconsistent with the message.

C. Filler words (―ums‖) are used

excessively.

D. Articulation and pronunciation

tend to be sloppy.

E. Over dependence on notes may

be observed.

54

SACS Standards Relating to Assessment

• 2.5 The institution engages in ongoing, integrated, and institution-wide research-based

planning and evaluation processes that incorporate a systematic review of programs and

services that (a) results in continuing improvement and (b) demonstrates that the

institution is effectively accomplishing its mission.

• 2.7.3 The institution requires in each undergraduate degree program the successful

completion of a general education component at the collegiate level that

(1) is a substantial component of each undergraduate degree

(2) ensures breadth of knowledge, and

(3) is based on a coherent rationale.

• 3.3.1 The institution identifies expected outcomes for its educational programs and its

administrative and educational support services; assesses whether it achieves these

outcomes; and provides evidence of improvement based on analysis of those results.

• 3.4.1 The institution demonstrates that each educational program for which academic

credit is awarded (a) is approved by the faculty and the administration, and (b) establishes

and evaluates program and learning outcomes.

• 3.4.12 The institution places primary responsibility for the content, quality, and

effectiveness of its curriculum with its faculty.

• 3.4.13 For each major in a degree program, the institution assigns responsibility for

program coordination, as well as for curriculum development and review, to persons

academically qualified in the field.

• 3.5.1 The institution identifies college-level competencies within the general education

core and provides evidence that graduates have attained those competencies.

• 3.7.2 The institution regularly evaluates the effectiveness of each faculty member in

accord with published criteria, regardless of contractual or tenured status.

• 3.7.3 The institution provides evidence of ongoing professional development of faculty as

teachers, scholars, and practitioners.

• 4.2 The institution maintains a curriculum that is directly related and appropriate to its

purpose and goals and to diplomas, certificates, or degrees awarded.

55

PDCCC Library Teaching Resources & Assessment Bibliography

Angelo, Thomas A. and K. Partricia Cross. Classroom Assessment

Techniques. San Francisco, CA: Jossey-Bass, 1993.

Blythe, Hal, and Charlie Sweet. It Works for Me!: Shared Tips for Teaching.

Stillwater, OK: New Forums, 1998.

Blythe, Hal, and Charlie Sweet. It Works for Me, Too!: More Shared Tips for

Effective Teaching. Stillwater, OK: New Forums, 2002.

Boylan, Hunter. What Works: Research-Based Best Practices in

Developmental Education. Boone, NC: Appalachian State U, 2002.

Cushman, Kathleen. First in the Family: Advice about College from First-

Generation Students; Your College Years. Providence, RI: Next

Generation, 2006.

D'Errico, Deanna, ed. Effective Teaching: A Guide for Community College

Instructors. Washington: The American Association of Community

Colleges, 2004.

Farnsworth, Kent, and Teresa Bevis. A Fieldbook for Community College

Online Instructors. Washington: Community College Press, 2006.

Friday, Bob. Create Your College Success : Activities and Exercises for

Students. Belmont, CA: Wadsworth, 1988.

Gabriel, Kathleen F. Teaching Unprepared Students: Strategies for

Promoting Success and Retention in Higher Education. Sterling, VA:

56

Stylus Publishing, 2008.

Gallien Jr., Louis B., and Marshalita S. Peterson. Instructing and Mentoring

the African American College Student: Strategies for Success in Higher

Education. Boston: Pearson, 2005.

Jewler, A. Jerome, John N. Gardner, and Mary-Jane McCarthy, eds. Concise

Edition. Your College Experience: Strategies for Success. Belmont, CA:

Wadsworth, 1993.

Holkeboer, Robert. Right from the Start : Managing Your Way to College

Success. Belmont, CA: Wadsworth, 1993.

Johnson, Elaine B. Contextual Teaching and Learning : What It Is and Why

It's Here to Stay. Thousand Oaks, CA: Corwin P, 2002.

Kanji, Gopal K. 100 Statistical Tests. 3rd ed. London: Sage, 2006.

Leamnson, Robert. Thinking About Teaching and Learning: Developing

Habits of Learning with First Year College and University Students.

Sterling, VA: Stylus, n.d.

Lieberg, Carolyn. Teaching Your First College Class: A Practical Guide for

New Faculty and Graduate Student Instructors. Sterling, VA: Stylus,

2008.

Linehan, Patricia. Win Them Over: Dynamic Techniques for College Adjuncts

and New Faculty. Madison, WI: Atwood Publishing, 2007.

Magnan, Robert. 147 Practical Tips for Teaching Professors. Madison, WI:

Atwood Publishing, 1990.

Mamchur, Carolyn. A Teacher's Guide to Cognitive Type Theory and Learning

Style. Alexandria: Association for Supervision & Curriculum

57

Development, 1996.

McGlynn, Angela P. Successful Beginnings for College Teaching: Engaging

Your Students from the First Day. Madison, WI: Atwood Publishing,

2001.

Nilson, Linda Burzotta. Teaching at Its Best : A Research-Based Resource for

College Instructors. 2nd Ed. Bolton, MA: Anker, 2003.

Palloff, Rena M. and Keith Pratt. The Virtual Student : A Profile and Guide to

Working with Online Learners. San Francisco: Jossey-Bass, 2003.

Palomba, Catherine A. and Trudy W. Banta. Assessment Essentials:

Planning, Implementing, and Improving Assessment in Higher

Education. San Francisco: Jossey-Bass, 1999.

Pregent, Richard. Charting Your Course: How to Prepare to Teach More

Effectively. Madison, WI: Atwood Publishing, 2000.

Roueche, John E. and Suanne D. Roueche. High Stakes, High Performance :

Making Remediial Education Work. Washington: Community College

Press, 1999.

Roueche, John E., Eileen E. Ely, and Suanne D. Roueche. In Pursuit of

Excellence: The Community College of Denver. Washington: Atwood,

2001.

Sarasin, Lynne C. Learning Style Perspectives : Impact in the Classroom.

Madison, WI: Atwood, 1999.

Schuh, John H., M. Lee Upcraft, et.al. Assessment Practice in Student

Affairs: An Applications Manual. San Francisco: Jossey-Bass, 2001.

Sims, Ronald R., and Serbrenia J. Sims, eds. The Importance of Learning

58

Styles : Understanding the Implications for Learning, Course Design

and Education. Westport, CT: Greenwood, 1995.

Stevens, Dannelle D. and Antonie J. Levi. Introduction to Rubrics: An

Assessment Tool to Save Grading Time, Convey Effective Feedback

and Promote Student Learning. Sterling, VA: Stylus Publishing, 2005.

Student Survival Guide. New York: College Entrance Exam Board, 1991.

Taylor, Terry. 100% Information Literacy Success. Clifton Park, NY:

Thomson, 2007.

Upcraft, M. Lee, John H. Schuh, and John H. Schuh. Assessment Practice in

Student Affairs : An Applications Manual. San Francisco: Jossey-Bass,

2000.

Vernoy, Mark, and Diana Kyle. Behavioral Statistics in Action. Boston: McGraw-Hill, 2002.

Walvoord, Barbara E. Assessment Clear and Simple: A Practical Guide for

Institutions, Departments and General Education. San Francisco:

Jossey-Bass, 2004.

Walvoord, Barbara E. and Virginia Johnson Anderson. Effective Grading: A

Tool for Learning and Assessment. San Francisco, CA: Jossey-Bass,

1998.

Weimer, Maryellen and Joan L. Parrett. How Am I Teaching? Forms and

Activities for Acquiring Instructional Input, Madison, WI: Atwood

Publishing, 2002.

59

Weimer, Maryellen and Rose Ann Neff. Teaching College: Collected

Reading’s for the New Instructor. Madison, WI: Atwood Publishing,

1998.

60

Internet Resources for Assessment General Principles of Assessment: http://www.tcc.edu/welcome/collegeadmin/OIE/SOA/principles.htm Writing Measurable Learning Outcomes: http://www.adprima.com/objectives.htm Evaluation Methods to Measure Outcomes by Programs: http://www.unf.edu/acadaffairs/IE/alc/ Types of Measures: http://www.provost.wisc.edu/assessment/manual/manual2.html Action Strategies to Closing the Loop: http://www.siue.edu/~deder/assess/catmain.html Rubric Creations: http://rubistar.4teachers.org/index.php , http://pareonline.net/getvn.asp?v=7&n=3 , http://imet.csus.edu/imet2/nicher/toohotwebquest/evaluation.html , http://pareonline.net/getvn.asp?v=7&n=25 , http://www.teach-nology.com/web_tools/rubrics/ Rubric Generator-Rubistar: http://rubistar.4teachers.org/ Towson Assessment Resources: http://pages.towson.edu/assessment/office_of_assessment.htm NC State Assessment Resources: http://www2.acs.ncsu.edu/UPA/assmt/resource.htm College of Du Page Resources: http://www.cod.edu/outcomes Assessment Peer Review Electronic Journal: http://PAREonline.net Virginia Assessment Group (VAG): http://virginiaassessment.org/RPAJournal.php National Council on Measurement in Education (NCME): http://ncme.org American Educational Research Association: http://aera.net Examples of Critical Thinking Scoring Rubrics: Holistic Critical Thinking Scoring Rubric: http://66.132.144.88/pdf_files/rubric.pdf Analytic Critical Thinking Rubric: http://www.neiu.edu/~neassess/pdf/CriThinkRoger-long.pdf

Writing Learning Objectives

Basic Guidelines (and Examples) for

http://www.mapnp.org/library/trng_dev/lrn_objs.htm

How do I write an instructional objective?

http://edtech.tennessee.edu/~bobannon/objectives.html

How to Write Clear Objectives

http://tlt.its.psu.edu/suggestions/research/Write_Objectives.shtml

How to Write Learning Objectives in Behavioral Form

http://www.adprima.com/objectives.htm

Understanding Objectives

http://edweb.sdsu.edu/courses/EDTEC540/objectives/ObjectivesHome.html

Guidelines for writing learning objectives in librarianship, information science and

archives administration

http://www.unesco.org/webworld/ramp/html/r8810e/r8810e00.htm#Contents

Quick Guide to Writing Learning Objectives

61

http://www.nwlink.com/~donclark/hrd/templates/objectivetool.html

Writing Learning Objectives

http://www.arl.org/training/ilcso/objectives.html

Writing good work objectives

http://home.att.net/~nickols/workobjs.htm

Writing instructional objectives: The what, why how and when.

http://www.sogc.org/conferences/pdfs/instructionalObj.PDF

Blooms Taxonomy

Affective Domain

http://www.itc.utk.edu/~jklittle/edsmrt521/affective.html

Assessing Learning Objectives Bloom's Taxonomy

http://www.ion.uillinois.edu/resources/tutorials/assessment/bloomtaxonomy.asp

Bloom‘s Taxonomy

http://www.officeport.com/edu/blooms.htm Cognitive Domain

http://www.itc.utk.edu/~jklittle/edsmrt521/cognitive.html

Psychomotor Domain

http://www.itc.utk.edu/~jklittle/edsmrt521/psychomotor.html

Instructional Design

http://carbon.cudenver.edu/~mryder/itc_data/idmodels.html#isd

Assessment Curriculum Development Performance Criteria

http://its.foxvalleytech.com/iss/curric-assessment/CRITCOND.html

How to Write an Assessment Based on an Objective

http://www.adprima.com/assessment.htm

Performance Criteria

http://its.foxvalleytech.com/iss/curric-assessment/CRITCOND.html

Multiple Choice Questions and Bloom‘s Taxonomy

http://web.uct.ac.za/projects/cbe/mcqman/mcqappc.html