SECTION ONE EXPLORATION OF THE STUDY AREA …

121
SECTION ONE EXPLORATION OF THE STUDY AREA 1.1 INTRODUCTION AND BACKGROUND In the previous South African education dispensation, prior to the democratic elections of 1994, educators were not given the opportunity to participate in the process of evaluating and selecting Learning and Teaching Support Materials (LTSM) for their classrooms (Gauteng Department of Education LTSM Policy 2002:1) This resulted in selection of inappropriate LTSM or materials that were not relevant to the school community. This was in violation of the educators’ right to having a voice in the selection of materials and making a contribution to education as stipulated in the United Nations High Commission for Human Rights (UNHCHR), article 26 of the Universal Declaration of Human Rights. This is particularly because parents as well as teachers have, through community- based forums, expressed their wish to be included in order to participate in the process (Monyokolo 1993:9). In South Africa, evaluation of Learning and Teaching Support Materials (LTSM) is an important function for every provincial education department. This includes the Gauteng Department of Education (GDE), which is the focus of this study. This study therefore will inquire into how educators as evaluators engage with the instrument that is used to evaluate LTSM. An endeavour is made to detect problems and successes that educators as evaluators experience in the process and whether such an evaluation instrument is effective, increases efficiency and has impact. In the new democratic dispensation in South Africa, educators are regarded as equal, active participants in the process of selecting LTSM. They are encouraged to work cooperatively with their peers. They are given the status of what Dobson (2000:206) refers to as internal evaluators of LTSM. The inclusion 1

Transcript of SECTION ONE EXPLORATION OF THE STUDY AREA …

Page 1: SECTION ONE EXPLORATION OF THE STUDY AREA …

SECTION ONE

EXPLORATION OF THE STUDY AREA

1.1 INTRODUCTION AND BACKGROUND

In the previous South African education dispensation, prior to the democratic

elections of 1994, educators were not given the opportunity to participate in the

process of evaluating and selecting Learning and Teaching Support Materials

(LTSM) for their classrooms (Gauteng Department of Education LTSM Policy

2002:1) This resulted in selection of inappropriate LTSM or materials that were

not relevant to the school community. This was in violation of the educators’ right

to having a voice in the selection of materials and making a contribution to

education as stipulated in the United Nations High Commission for Human

Rights (UNHCHR), article 26 of the Universal Declaration of Human Rights. This

is particularly because parents as well as teachers have, through community-

based forums, expressed their wish to be included in order to participate in the

process (Monyokolo 1993:9).

In South Africa, evaluation of Learning and Teaching Support Materials (LTSM)

is an important function for every provincial education department. This

includes the Gauteng Department of Education (GDE), which is the focus of this

study. This study therefore will inquire into how educators as evaluators engage

with the instrument that is used to evaluate LTSM. An endeavour is made to

detect problems and successes that educators as evaluators experience in the

process and whether such an evaluation instrument is effective, increases

efficiency and has impact.

In the new democratic dispensation in South Africa, educators are regarded as

equal, active participants in the process of selecting LTSM. They are

encouraged to work cooperatively with their peers. They are given the status of

what Dobson (2000:206) refers to as internal evaluators of LTSM. The inclusion

1

Page 2: SECTION ONE EXPLORATION OF THE STUDY AREA …

of educators in this process is also in line with democracy as borne in the

National Curriculum Statements (RNCS) and the South African Schools Act of

1996. This is also an endeavour to fulfil the obligation of the GDE to ensure that

materials are carefully and professionally selected for learners and educators, as

stated in the Department of Education (DoE) policy document (2002:3).

In the GDE, the process discussed above is managed by the LTSM Sub-

directorate. Ideally the LTSM evaluation and selection process should take

place 18 months prior to the date scheduled for the delivery of LTSM to schools.

This is particularly important because the intention of the GDE is that all learners

will have LTSM by the first day of the school year.

To achieve this obligation, the GDE designed and utilised a National Curriculum

Statement compliant evaluation instrument for the Foundation and Intermediate

Phase (See Addendum A). As stated in the foreword of the GDE evaluation

instrument (2003:3), the instrument is used on a continuous basis for evaluating

LTSM at various levels (e.g. Provincial, school, phase and grade levels). The

criteria stipulated in the evaluation instrument and their utilization are what the

GDE demands from educators and other stakeholders and these may be

referred to as professional demands or bureaucratic conceptions (Scriven in

Kellaghan, Stufflebeam and Wingate 2003:17). Furthermore, there is an

endeavour to give the educators an opportunity to test for the validity of the

evaluation instrument and to make value judgments. This is in compliance with

what Scriven (in Kellaghan et al., 2003:16) classifies as normative theories.

The evaluation instrument is Phase and Grade specific. The front cover of the

instrument enquires about the evaluators who will actively participate in the

process of LTSM evaluation, details on the LTSM (e.g. Title, grade, ISBN,

publisher) and the identification of the instrument (its name and Phase/ Grade).

The first page of the instrument is a Foreword by the GDE. In this section, an

attempt is made to introduce the process and procedures to be followed during

evaluation as well as the intended product, the Review Guide or catalogue.

2

Page 3: SECTION ONE EXPLORATION OF THE STUDY AREA …

Coupled with the introduction to procedure to evaluate, are the notes to the

evaluator as well as the lay out of the sections and divisions of the actual

instrument.

As LTSM head in the GDE, I am involved in the evaluation and selection of

Learning and Teaching Support Materials (LTSM) for the schools. The task of

LTSM evaluation outlined above is accompanied by activities such as the training

of educators on the same and related tasks, appointment of tenderers and

service providers and procurement of LTSM for the schools. As the project

manager, I have always been concerned about the effectiveness of the

instrument used for LTSM evaluation.

1.2 STATEMENT OF THE PROBLEM

The impracticality of evaluators of materials performing the task of LTSM

evaluation without a relevant instrument is a problem. To attend to this problem,

the GDE responded by providing an evaluation instrument to be used by

evaluators at different levels. However, the main concern should be whether or

not the instrument addresses the needs or reaches the intended outcomes. It is

also important to remember that the main intended outcome is for the educator

to be able to evaluate and select appropriate LTSM that are usable in the

classroom.

The above statement is an indication of the reasons for this inquiry to take place.

This study intends to guide educators as evaluators on how to engage with the

evaluation instrument. This will, therefore, assist in establishing the usefulness

of the evaluation instrument. The findings of this study could assist the GDE in

deciding whether the evaluation instrument they are presently using should be

adapted, adjusted, maintained or completely discarded and replaced by

something else.

3

Page 4: SECTION ONE EXPLORATION OF THE STUDY AREA …

Based on this background, the research question for this study is: Does theGDE instrument for LTSM evaluation increase effectiveness, efficiency andimpact in the evaluation, selection and utilization of materials?

1.3 AIMS OF THE STUDY

The main aim of the study is to find whether the GDE evaluation instrument

effectively makes an impact on the LTSM selection process, which could

therefore lead to efficiency among all the participants.

The sub-aims of the study include knowing if:

• evaluators can use the instrument independently

• the educators are willing to do evaluation as part of their daily routine

• the instrument encourages collaboration among educators

• the instrument is regarded as a core element for LTSM evaluation

• The instrument does empower educators to select only appropriate

LTSM.

1.4 RESEARCH DESIGN AND METHODOLOGY

To make this inquiry possible, the researcher adopts a predominantly qualitative

approach with a minimum slant toward quantitative research. In qualitative

research, as opposed to quantitative, the researcher is able to obtain more detail

and a larger variety of perspectives to the issues that are being investigated.

Henning, Van Rensburg and Smit (2004:6), describe this participative approach

as a more open-ended way in which the participants give their opinions, view

points and also demonstrate their actions. In other words, the variables are not

controlled. Melody (2002:148) emphasises that the participants are equally

human and should be given that status during inquiry. Furthermore, this type of

research allows all the participants (researcher and respondents) the liberty to

probe issues deeper as discussions progress.

4

Page 5: SECTION ONE EXPLORATION OF THE STUDY AREA …

In quantitative research the variables are controlled. The instruments are

designed to control the variables, for instance the respondents are not allowed to

express their viewpoints over and above what is contained in the questions on

the research instruments. Apart from causing frustration among the

respondents, the research inquiry becomes limited by what is contained in the

instruments. To avoid this situation, in this inquiry the researcher uses both

qualitative and quantitative to enable the respondents to participate in

discussions and clarify their respondents or even significant information that may

have been omitted. The research methodology, however, attempts to allow for

looking at data, finding the meaning of data, interpreting the meaning of data and

applying data in the best suitable manner (Henning et al.,, Van Rensburg and

Smit, 2004:4).

Furthermore, in the following section the literature on materials evaluation is

reviewed. Thereafter, selection of a specific research methodology is made with

the elaboration on the way data will be analysed resulting in a discussion of the

findings. A literature review gives the theoretical background in which various

authors and specialists give their perspective on the evaluation of learning

materials, and relevant theories such as radical and social constructivism and

transformative learning. In the research methodology section, there is

elaboration on the data, identification of the population and sample as well as

criteria for selection of materials’ evaluators.

In addition, I analyse the context within which the research takes place and also

described the sampling method I used. However, this section would not be

complete without the discussion on data collection methods in which I use survey

questionnaires and focus group interviews. I also give an exposition of how I

engage in the exercise to measure whether or not the questionnaire is valid and

reliable. The findings of the research are only useful if they are analysed and

made ready for implementation purposes. To analyse the data, I employ the

constant comparative method as presented by Maykut and Morehouse (1994).

5

Page 6: SECTION ONE EXPLORATION OF THE STUDY AREA …

The data collection methods used are survey questionnaires and focus group

interviews (considering their validity and reliability). After that, I present the

findings and perform data analysis to back up my conclusion and make

recommendations.

1.5 CONCEPT CLARIFICATION

The above-stated research design and methodology, in this inquiry, are

implemented to assist the GDE in the possible best way to evaluate LTSM.

Later on, in this section, the rationale for the inquiry is explained. However, prior

to delving into the rationale and theoretical orientation, it is important to clarify

what LTSM are and also to explore their different categories and types. The

National Department of Education (DoE) (2002:3), in its LTSM policy document,

adopted the term “Learning and Teaching Support Materials” (LTSM). Prior to

that, the term had been “Learning or Learner Support Materials” (LSM). The

two-fold naming of LSM was based on the fact that Departmental officials argued

whether it should be ‘Learner’ or ‘Learning’ Support Materials. Some officials

thought because it is intended to support the learner, it should carry the prefix

‘Learner’ (GDE 2003:3).

In addition, policy also stipulates that LSM was intended to support the learning

process undergone by the child, and should therefore carry the prefix ‘Learning’

(GDE 2003:3). This poses a challenge to the social constructivists who view

learning as a social collaborative process which happens between an individual

learner and others (co-learners or community). This implies that learning

happens in a dynamic social context (Gravett 2001:20). However, what is of

essence is that the learner actively constructs knowledge in the learning process,

in which his/her existing knowledge and experience play a crucial role. This

implies that the acronym LSM can harmlessly be utilized for both the concepts.

6

Page 7: SECTION ONE EXPLORATION OF THE STUDY AREA …

The GDE and Gauteng Institution for Educational Development (GIED) (1999:2)

stated that the term ‘Learning Support Materials’ refers to any kind of materials

that can be used to assist the learner to learn or facilitates learning. ‘Teaching

Support Materials’ (TSM) refers to materials that assist the teacher to teach. The

justification for the inclusion of the name ‘teachers’ in the term LTSM was that

teachers felt they were neglected and not given much support in terms of

teaching materials. The DoE decided to allocate some funds for Teacher

Support Materials. For the purpose of this study, I will use the collective term

‘LTSM’, since it encompasses both TSM and LSM.

The collective LTSM includes printed media, stationery and other materials.

Printed media include, among other things, textbooks, educators’ guides,

learners’ guides, learner workbooks, readers, atlases, dictionaries, magazines,

newspapers, charts and posters. Stationery includes exercise books, drawing

books, examination books, map books, paper glue, crayons, pencils, pens, fibre

pens, marking pens, news print, duplicating paper (not for office but for learners)

and erasers. ‘Other Materials’ include materials falling outside the

aforementioned categorization like educational equipment such as science kits,

puzzles, laboratory equipment, veniers, scales, tape measures, puzzles and

games. It also includes consumables such as soap, groceries for Home

Economics, some cartridges and glue (GIED/GDE 1999:2).

The above clarification is made to alert the reader that LTSM is not limited to

textbooks only and that an evaluation instrument needs to have sections for

evaluating a wide variety of LTSM. Other types of LTSM, such as educators’

guides, learners’ guides and workbooks, may be regarded as forms of textbooks.

As specified by Davies (2000:126), textbooks, in whatever form, are aimed at

fulfilling three main functions, namely ‘interpersonal’ (relationship between writer

and reader), ‘ideational’ (selection of information) and ‘textual’ (constructing a

coherent message). By means of appropriate instruments, evaluators

endeavour to establish whether these functions are fulfilled by LTSM.

7

Page 8: SECTION ONE EXPLORATION OF THE STUDY AREA …

1.6 OUTLINE OF THE REPORT

This work consists of five sections each dealing with a particular aspect. The

outline of the sections is as follows:

Section One introduces the research project to the reader. The introduction

provides a background in which the reader is given the context and focus of the

study, in particular the history of evaluation and selection in the GDE (that is

located in South Africa). The problem that necessitates this research is stated,

followed by the aim of the study. There is also a description of the research

design and method. With regard to concept clarification, the main terms that are

used in the essay are clarified.

Section Two consists of a Literature Review, in which is explored some of the

texts on evaluation and selection of materials, and evaluation of project

effectiveness, efficiency and impact. The first part deals with the different views

on definition while the second part is on effectiveness, efficiency and impact of

projects. This is followed by an exploration of reasons for evaluation to take

place. Different evaluation models, as presented by Boulmetics and Dutwin

(2000) are discussed.

Section Three first discusses research methodology in the form of population,

sample and data collected for analysis. This is coupled with research orientation

and context in which the scene is set. The survey questionnaire (including test

for validity and reliability) and focus group interviews that I used to collect data

are fully discussed. This discussion is not complete without the description of

the interview schedule,

Section Four mainly focuses on data analysis and research findings. Actual

data analysis takes place. The findings are discussed, grouped into themes

(using Maykut & Morehouse, 1994:127), and a constant comparative method is

presented. These findings are matched against the reviewed literature.

8

Page 9: SECTION ONE EXPLORATION OF THE STUDY AREA …

Section Five concludes the study and recommendations are made. In this

section, conclusions are drawn based on the findings. Finally, recommendations

are made to the GDE on areas that they might need to make improvements on.

1.7 SUMMARY

In this section, the overall intention of this inquiry was introduced. I explored the

different definitions of LTSM, the key concept in this inquiry. Furthermore, I gave

the background and history of the evaluation of LTSM starting from a general

perspective and narrowed down to the GDE context. This is because what

happens in the world directly has a bearing on what happens in a small area or

locality. In the last parts of this section, I state the problem which is a focus of

this study. I would like to re-iterate the aim of this inquiry which is to evaluate the

effectiveness, efficiency and impact of the evaluation instrument used by the

GDE to test whether LTSM are appropriate for use in the classroom. The next

section focuses on the Literature Review which deals with the exposition of what

various texts on evaluation specify.

SECTION TWO

9

Page 10: SECTION ONE EXPLORATION OF THE STUDY AREA …

LITERATURE REVIEW

2.1 INTRODUCTION AND DEFINITION OF THE CONCEPT EVALUATION

In the previous section, it has been argued that the prominent aspect on which

this work is based is the evaluation of an evaluation instrument. This therefore

suggests that prior to engaging with the actual exercise, it is mandatory to

explore the works of different writers and theorists on basic evaluation. In so

doing, I shall, at all times endeavour to apply the knowledge in the exercise of

evaluating the instrument used by GDE to evaluate LTSM. Some of the

knowledge needs to be considered prior, during and after an evaluation session

or project. It should also be noted that evaluation is a process that requires one

to construct, test and utilize certain special artefacts, for instance an evaluation

instrument. Evaluation exercises, as suggested by Boulmetics and Dutwin

(2000:2) happen mostly in controlled environments.

This section of my work consists of four sub-sections that deal with literature

review in some detail. In Sub-section 2.1, I introduce and define the generic

concept of evaluation and later on attempt to narrow it down specifically to the

evaluation of LTSM. Definitions by various educationists and evaluation

theorists are explored and finally fused to form an understanding of evaluation of

LTSM. In sub-section 2.2, I analyze deeply three main concepts that are

centered on evaluation, namely: efficiency, effectiveness and impact. These

concepts also serve the primary basis for conducting an evaluation. Hence, they

connect this sub-section with the next one (Sub-section 2.3) which specifically

deals with formal reasons for evaluation as put by some evaluation experts.

Section 2.4 explores these evaluation models: discrepancy, goal-free,

transaction, decision and goal-based model and the criteria used for evaluating

LTSM. Such criteria are primarily to test whether standards are adhered to.

10

Page 11: SECTION ONE EXPLORATION OF THE STUDY AREA …

These standards include compliance with the curriculum requirements (GDE

LOTEA policy 2003: 2).

In view of the above, the starting point will be to explore the definition of the

concept ‘evaluation’. Boulmetics and Dutwin (2000:4) define evaluation as a

systemic process of collecting and analyzing data in order to determine whether

and to what degree objectives are achieved. Furthermore, Boulmetics and

Dutwin (2000:4) view evaluation as a systemic process of collecting and

analyzing data in order to make informed decisions for future use. Evaluation is

perceived as a social process that is different from development and which

might, if not carefully employed lead to a ‘’we- they’’ polarity. It is therefore upon

education departments to ensure that none of the stakeholders feel excluded

from the process. (Burtman & Fletcher, in Borich 1974:39).

It is quite evident that an education department that strives for unity, like the

GDE, will at all costs avoid a situation that may arise with a ‘we-they’ polarity. A

department would like its employees to feel like they are part of the same

homogeneous group. In that set up, evaluation forms part of what may also be

described as identification of discrepancies between where a project is and

where its leaders would like to be (an indicator for the real and ideal situations).

Furthermore, a project should have the qualities of engagingness, accessibility

and generalizability. These constructs are also critical in understanding the

support that texts must provide children (Menton & Hiebert 1999: 07).

Education departments need to have measures to ensure the absolute objectivity

of evaluation. This would assist in producing results that are acceptable to

publishers who are stakeholders in education and more importantly, businesses,

who need to sell their LTSM (Denning (1992:14). To be more precise, the

evaluation of LTSM entails objective and intensive reading, bearing in mind

criteria as stipulated in an evaluation instrument; writing down analytic points

(strong and weak) to determine whether the LTSM are educationally sound,

11

Page 12: SECTION ONE EXPLORATION OF THE STUDY AREA …

scoring and writing qualitative reports/ reviews about every component of a

publisher’s LTSM (GDE 2003: 1).

These definitions suggest that when an organization engages in an evaluation

the primary purpose is to see rapid change, growth and development. Borich

(1974:27) juxtaposes educational evaluation with the cost of living stating that

both have one thing in common, namely a propensity for rapid change. The

GDE therefore, needs to be dynamic to keep up with the demands of the

publishing world and the curriculum.

The concept ‘evaluate’ together with other related or synonymous concepts

have a common goal. This refers to words like ‘assess’, ‘appraise’, ‘weigh’,

‘consider’,’ review’ ‘gauge’, ‘calculate’ ‘moderate’, ‘adjudicate’ ‘referee’ ‘umpire’

and ‘estimate’. The common goal or intention implied in the aforementioned

terms is to find the value, worth or usefulness of a subject. This is what

evaluators of materials are conscious of when evaluating LTSM.

Based on the above definitions, one can assert that an effective evaluation

process is one that operates within certain parameters or criteria and its planning

is based on a specific purpose, intended outcomes and objectives. Furthermore,

the process of evaluation is scientific and is driven by data collection and data

analysis. The reasons for the effort of collecting and analyzing data are different

and reflect a notable difference in philosophies (Boulmetics & Dutwin 2000:4).

Evaluation, review and selection of LTSM are interdependent processes and

serve as a pre-requisite for LTSM procurement. Depending on a number of

possible causative factors, the GDE possesses the right to use its prerogative in

deciding whether or not to undertake any one of the three steps. However, the

definition below is meant for the purpose of indicating the common

understanding of the concepts among all users and participants within the GDE

(GDE LOETA policy 2003:3).

12

Page 13: SECTION ONE EXPLORATION OF THE STUDY AREA …

For the purpose of this inquiry, data was collected to examine the effectiveness

of the evaluation instrument used by GDE to evaluate LTSM. The primary focus

in this exercise is establishing the veracity of the instrument so as to see if it is

worth retaining as is or needs to be adjusted to meet the requirements of the

GDE. This implies that evaluation makes a claim of the value of the instrument

in relation to the overall process of LTSM evaluation. As Boulmetics and Dutwin

(2000:5) note:

Many experts agree that an evaluation should only assess program results

(whether objective and intended outcomes were met) but also identify ways

to improve the program

The evaluator, a human being, is central to the evaluation process. Project

leaders, according to Borich (1974:27), should always bear it in mind that one of

the most significant developments in the intended rapid growth referred to above

is the human and personal nature of the evaluator. The evaluator deals with a

variety of people like other evaluators, publishers, colleagues and project

leaders. In addition to these relationships, the evaluator also needs to engage

with his/her artefacts in order to be able to measure the LTSM successfully.

The role of the evaluator, in this context is for good and not for ill – to protect the

end users from purchasing inappropriate LTSM and to assist the publishers on

improving their LTSM to the required standards. Borich (1974:101) summarizes

the role of the evaluators as: to demonstrate that the materials developer has

been successful; that the materials meet the requirements and intended

outcomes; and that the overall quality of the materials is to the required standard.

One may therefore conclude that the objective for LTSM Evaluation and Review

is to make a positive impact on the effectiveness and efficiency of its

provisioning, defined as having the right materials at the least cost in the right

institutions at the right time (Eastern Cape Department of Education, 2004).

13

Page 14: SECTION ONE EXPLORATION OF THE STUDY AREA …

In order to perform the task of evaluation of LTSM successfully, the evaluator

needs to be given preset criteria so as to ascertain the parameters within which

to function. The criteria are set by the relevant education department, in this

regard the Gauteng Department of Education (GDE) through its LTSM unit. The

criteria are enumerated in a grade or phase – specific evaluation instrument that

is adapted to the needs of the curriculum, the Revised National Curriculum

Statement (GDE 2003:2).

Upon completion of the evaluation process, education departments select LTSM

that meet specified minimum requirements and place those in a list of approved

LTSM. The list is eventually converted into a catalogue which the GDE names

the ‘List of Approved Materials’. It is important to state that evaluation occurs at

different levels: outside the school at the Evaluation Centre and at the school by

individuals and teacher teams (e.g. LTSM committees within the school). This

implies that schools have a role to play in the selection of LTSM considering the

needs of their teachers, learners and the surrounding school communities (GDE

2003:2).

However, it is important to note that evaluation of LTSM is not the only activity in

an education departments’ process map (see Addendum C) but forms an

important and indispensable area. This is evident in Figure 1 and 2 as specified

by models used by the Eastern Cape Department of Education and GDE

respectively (GDE 2003:3).

2.2 PERFORMANCE OF EVALUATION TO MEASURE EFFICIENCY,EFFECTIVENESS AND IMPACT

Public and private enterprises are faced with the task to lead, monitor and control

projects and programmes. Likewise, the GDE is also faced with such a task.

The LTSM evaluation and selection process is one of the biggest projects within

the GDE. In ensuring that the project succeeds and is sustainable, a project

leader needs to be a good decision maker. This entails assessing the project’s

14

Page 15: SECTION ONE EXPLORATION OF THE STUDY AREA …

efficiency, effectiveness and impact. If the results prove to be negative, the

project leader should consider either abandoning the project or introducing new

alternative justifiable measures to rescue the project.

To achieve the aforementioned, as reflected on its LTSM evaluation policy

known as Limited Open Equivalent to Approval (LOETA) (Addendum E), the

GDE shows its participative approach and has certain demands towards its

officials, educators, evaluators and learners concerning criteria and standards for

LTSM evaluation and selection. To accomplish that, criteria for evaluation and

selection are laid down and expressed through the evaluation instrument. This is

what the GDE does not negotiate with its stakeholders but states as a

requirement. Scriven (in Kellaghan, Stufflebeam & Wingate 2003:17) refers to

similar requirements as professional demands or bureaucratic conceptions.

While employing the above, it is however important to note that an education

department should not intend to use professional demands to limit the educators’

mobility or liberality but should aim to display a pattern and to set parameters

within which they should work to benefit the process of teaching and learning.

To meet this end, in this study, there is an endeavour to give the educators an

opportunity to test for validity of the evaluation instrument and to do value

judgment. This is in compliance with what Scriven (in Kellaghan et al., 2003:16)

classifies as Normative theories. The implication is that what evaluators do is

driven by the current swings of the different sections of society or stakeholders

like the public, their profession and their bureaucratic conceptions and

convictions. It could be outcome-oriented evaluation. In other words when

impact is calculated the outcomes are considered.

A professional evaluator, according to Scriven (in Kellaghan et al., 2003: 17), is

expected to do more than just reporting facts as they are. He/she should also

critique the findings and synthesize the results to measure the usefulness of the

process and of the instrument that is used to evaluate LTSM. By so doing an

15

Page 16: SECTION ONE EXPLORATION OF THE STUDY AREA …

evaluators’ team will end up with an evaluative conclusion about the merit of the

evaluation programme and as an outcome of the evaluation of LTSM.

The professional demands of the GDE suggest that the evaluator is bound by

the predetermined standards as decided by the curriculum and reflected on the

instrument for evaluation. For effectiveness, efficiency and impact, the way in

which the GDE conducts, this process should neither be carefree nor goal free.

Scriven (in Borich 1974:30) describes goal-free evaluation (GFE) as the one in

which the evaluator performs the function from a more general perspective.

Although the GDE would allow the evaluator to contribute existing knowledge

and skills, the evaluator is expected to follow the guidelines as specified on the

instrument for evaluation. The question for this inquiry is whether the instrument

is reliable and valid.

As outlined in section one of this work, educators and other stakeholders are

given the status of equal partners in education. This means that if changes are

to be implemented in the evaluation system, it should make sense to them. For

instance, a teacher and his/her class are a sub-system within a school but

changes cannot be forced upon them or else the changes will be resisted. This

implies that for the GDE to implement changes, the professional, the educator

needs to be consulted more particularly in the planning stages of the evaluation

process.

Effectiveness can also be measured against the selected LTSM themselves.

Serguin (1989:51) states that learning materials, in particular school textbooks,

are one of the factors that determine the effectiveness of teaching and learning

in school and the yield of educational systems. This implies that the quality and

relevance to the intended outcomes of LTSM are of significance. These qualities

should be carried by manuscripts that are presented to the education department

for evaluation. Content, pedagogical approach, language and illustration are

some of the determinant factors for textbook effectiveness (Potenza in Kromberg

et al., 1993:48).

16

Page 17: SECTION ONE EXPLORATION OF THE STUDY AREA …

The aforementioned still implies that educators are involved in a dialogic

construction of knowledge with their fellow educators (Gravett 2001:20). They

engage in discussions on the instrument and work in teams at the Evaluation

Centres and back at school as they have to co-operate with their School LTSM

Committees that the GDE established in the years 1997/8. Lincoln (2003:51)

refers to this knowledge as a socially constructed process. This is because the

educators regard themselves as a community of ‘knowers’. The professional

demands lead them to engage in participatory work. Therefore, in this work they

are referred to as ‘participatory evaluators’.

Boulmetics and Dutwin (2000:5) collectively term the concepts of efficiency,

effectiveness and impact as levels of programme evaluation. Efficiency refers to

the analysis of the costs. Costs may manifest themselves in rand amounts, the

number and variety of people involved in the project, time spent on the project,

facilities required for the project to run efficiently and materials or resources. An

evaluation instrument needs to measure whether processes are sufficient, cost

effective and correctly sequenced. It should also test whether system

arrangements support effective and efficient supply and prevent leakages

(Eastern Cape Education 2004: 17).

Effectiveness examines whether the planned project activities do what they were

supposed to do. In the case of the GDE, during LTSM evaluation, a scene is set

where there are teams of evaluators who hope to achieve selection of

appropriate LTSM as their intended outcome. Effectiveness would measure

whether the instrument does enable the evaluators to competently select only

appropriate LTSM. In other words, effectiveness is measured in substantive

changes in knowledge, skills, and attitudes. It is a way to assess whether the

activities provide the skills to run the project now and in future.

Impact is the most important evaluation consideration as it refers to attainment

and non-attainment of outcomes (Borich 1974:88) It examines whether there are

17

Page 18: SECTION ONE EXPLORATION OF THE STUDY AREA …

long-term and sustained changes (the extent) in a particular target population. In

the GDE evaluation process, if there is an impact, it will manifest itself as a

change among the evaluators and also within the evaluation process itself (the

decision making process whether to accept or reject a title that is being

evaluated). This study further sets out to find whether there are any desired

changes and whether the introduction of the evaluation instrument changed the

lives of evaluators and if it increases satisfaction among evaluators or not.

Based on the above, one may argue that the GDE is faced with the task of

measuring the impact of its LTSM evaluation project. This is purely because the

task of evaluating LTSM is extremely significant as it is a means to ensure that

the design features of the Revised National Curriculum Statement (viz: Critical

Outcomes, Assessment Standards Learning Outcomes) are achieved. This

would assist in ensuring that South Africa ends up with competitive, world-class

citizens.

To complement the above Boulmetics and Dutwin (2000:7) bring in another

perspective as they assert that the challenge to assessing impact is that it

manifests itself over time. The project managers may have resigned from their

employment when the results of impact are observed. My view is that this

statement is not completely true as part of the impact can be observed within a

short period. For instance, upon orientating evaluators, in a matter of hours they

commence with the evaluation activities. It is however difficult to measure

whether they are fully enabled by the orientation. As a preparation for the

assessment of impact, data may be collected by observing learner turnover – in

this case the evaluators’ turnover; circumstances at the schools after evaluation

orientation and activities and also by measuring change and how resources are

utilized. This is done to enable the project managers to identify areas of

improvement.

2.3 FORMAL METHODS FOR EVALUATING LTSM

18

Page 19: SECTION ONE EXPLORATION OF THE STUDY AREA …

The aforementioned suggests that evaluation is a specialized complex field of

work that may be accompanied by specific activities. These qualities of the

evaluation process suggest that in order to perform a reliable valid evaluation

activity, evaluators should be carefully selected and orientated and the

instrument for evaluation has to be developed in a manner that will enable it to

have as minimal error as possible. As implied by Brandes (in Borich I974:88),

the most important determinants of the criteria, depth and quality of the

evaluation instrument are the reasons to evaluate.

The implication from the above is also that In order to accomplish the task of

evaluation, provincial education departments have to design a criteria-bound and

curriculum compliant evaluation instruments. In view of the narrow gap between

the grades, the GDE designed an inclusive Foundation Phase and Intermediate

Phase evaluation instrument. However, due to complexity in the subsequent

Phases and grades and also the implementation plan of the national Department

of Education, a grade specific instrument may have to be designed and

implemented per grade. This has been the case with instruments that were

designed for curricula prior to the RNCS (GDE 2003:4).

Public and private organizations, including the GDE, invest millions of rands in

programmes and projects. This implies that they have a remarkable

accountability to sponsors, stakeholders and other affected parties. Therefore,

for fiscal purposes, project leaders have to conduct and monitor an evaluation

process at certain points in the development of a project. As suggested by

Boulmetics and Dutwin (2000:10), if an evaluation report communicates results

that indicate success, it could assist the project leaders to obtain funding for the

current and forthcoming year(s). In that case we may assert that the mandate

for evaluation is for fiscal purposes or that receiving the second year’s funding

may be based on the results of the evaluation. It is also for when a GDE LTSM

Head presents evaluation results, which could assist the senior managers on

deciding on the worth of the programme based on its successes and failures.

19

Page 20: SECTION ONE EXPLORATION OF THE STUDY AREA …

As noted, above, the success of the evaluation process is determined by its

participatory nature. This is evident in the way the GDE models its evaluation

process. As implied above, educators at the Evaluation Centre are grouped in

teams of three. This implies that an individual evaluator will first evaluate a

LTSM submission on his/ her own but later on, the team consolidates its reports

into a composite document.

At school level, a teacher works collaboratively with fellow Learning Area/

Programme educators and later on they communicate their decisions to the

School Learning and Teaching Support Materials Committee (SLTSMC). The

SLTSMC facilitates LTSM management at school level which needs demands a

great deal of time (Monyokolo 1993: 16). The results at the Evaluation Centre

serve as the provincial GDE decision on what should be used in the classrooms.

The result at school level informs the educators as to what kind and which

materials the educators desire to use in their classrooms.

The above-stated approach is not only encouraged at Evaluation Centres but

also at other evaluation sessions that take place at the schools. These involve

the SLTSMC, Learning Area/ programme committees and grade educators.

More interesting is that evaluators have also expressed the difficulties they face

as they engage in the evaluation process irrespective of their working teams and

consulting with one another.

The above-named difficulties have been communicated to officials at Head office

and District offices. This is done by means of a variety of ways, including

telephone calls and direct personal contact with LTSM units. By citing the

problems the evaluators were sending a message that they required some kind

of support from the GDE as they felt this was lacking.

The foregoing scenario attests to the notion that another formal reason for

evaluating is to justify a project or programme. The project leader may have

brought the idea to the department or organization he is working for and will also

20

Page 21: SECTION ONE EXPLORATION OF THE STUDY AREA …

be conscious that he or she is always under the scrutiny of other administrators,

managers, parents, teachers and publishers. These stakeholders may only

desire to communicate support provided the evaluation results are backed by

statistics.

In support of the above, Boulmetics and Dutwin (2000: 11) argue that:

Gut feelings, perceptions, innuendo, and anecdotes are comforting but

they are not convincing to who require more objective evidence. Even

data are not always effective in convincing people, but data at least act as

a common currency to demonstrate the value of your case.’ (Boulmetics &

Dutwin 2000:11). Furthermore, one may expect a good manager to

collect data on a continuous basis in anticipation in order to justify the

need when the time comes.

The GDE could justify their evaluation process by the fact that educators as

evaluators of LTSM consider the context of the school community. Thus their

generic selection criteria would almost be the same for electronic as it is for

printed materials. This would reflect utilization of technologies to support the

processes (Eastern Cape 2003: 17).

Stufflebeam (in Kellaghan et al., 2003:31) presents Context evaluation as the

first step in his Context Input, Process and Product Context Input, Process and

Product (CIPP) model of evaluation. In this model, ‘Input’ applies to the amount

of academic and professional skills, knowledge and experience that the

educators/ evaluators are expected to put in the evaluation of LTSM. Through

their experience evaluators are able to make their informed contribution and

assist in making reliable assumptions and conclusions that are influenced by the

evaluation process.

Mezirow (2000:17) refers to the above-stated assumptions as habits of mind.

The habits of mind present themselves in a variety of ways: by sociolinguistic

21

Page 22: SECTION ONE EXPLORATION OF THE STUDY AREA …

(cultural or ideological) means, moral-ethically (conscience), epistemic (learning

styles), philosophically (religious or philosophical doctrines), psychologically (self

concept and personality traits), and aesthetically (values and attitudes). These

varieties also influence an evaluator prior to taking a decision whether the

materials they evaluate are appropriate and may be utilized in the classroom. In

support, McKinney (2005: 5) supports that by stating that textbooks act as an

interface between the officially state-adopted and sanctioned knowledge.

The GDE is presently focused on evaluation of printed materials. In future, the

process will change as the focus shifts to evaluation of electronic media. At the

time of writing this study, this was imminent, as the GDE had begun installing

computer hardware in its schools. Once the infrastructure has been installed the

next step would be the evaluation and selection of software. The product is

almost the same in the sense that both evaluations will end up with a Review

Guide (GDE LOETA policy 2003: 1).

The justification reason connects to the next one, which is comparison reason.

In that respect, for comparison reasons, the project participants consider how

their project compares with other similar projects that are running at present or in

the past. That is another way to justify the project. The project leaders have to

be objective and try to consider if there are any other projects they would have

done instead of the current. This implies that the project leaders would also

need to audit the project needs and consider whether they are justifiable as

viewed by clients, colleagues, sponsors and other stakeholders.

This also means having a checklist to ascertain whether the project has or will

achieve the intended outcomes set out at its inception. To be more specific, one

may explain an outcome as what happens as a direct result of an action – this

could include training, service, teaching, orientation, work shopping. Boulmetics

and Dutwin (2000:25) argue that an evaluation takes place to determine whether

an action was effective or not. More interesting an outcome is viewed not only

as what was anticipated but also what may come as a surprise (a surprise may

22

Page 23: SECTION ONE EXPLORATION OF THE STUDY AREA …

either be positive or negative in nature). Boulmetics and Dutwin (2000:25) refer

to these as anticipated and unanticipated outcomes. If an outcome is achieved

on a longer range it may be termed ‘impact’.

The factors discussed above suggest that evaluations are done for decision-

making purposes. Every organization needs to make plans and policies.

Evaluations assist managers in deciding whether an area under a particular

project is worth adopting as policy or not. Managers use evaluation results to

distinguish between best-working programmes and those that need to be

discarded. This inquiry may also assist planners of evaluation of LTSM in

deciding whether the project is sustainable. Furthermore, evaluation is

performed to expose all the participants to new opportunities and new solutions

to current problems.

2.4 CHOOSING A MODEL FOR EVALUATION

People involved with evaluation of programmes are faced with a large spectrum

of evaluation models which they may or may not adopt to finalize their evaluation

exercise. More often a combination of more than one model is used for a given

evaluation purpose. In this section, I will discuss some of the popular models of

evaluation as suggested by Boulmetics and Dutwin (2000) and backed by the

views and contributions of other professionals.

2.4.1 Discrepancy Evaluation Model

Developed by Provus (in Boulmetics and Dutwin 2000:5), the Discrepancy Model

is based on the premise that a project is not in a vacuum but falls within an

organizational structure. The aim is not to determine the cause-and effect

relationship but to understand the evidence that is found based on an evaluation

exercise. The emphasis is on why events took place and not on whether or not

they occurred.

23

Page 24: SECTION ONE EXPLORATION OF THE STUDY AREA …

This model’s development takes place through stages of design, installation,

process, product, and cost-benefit analysis. There is a set of standards put to

examine performance. This is in line with what Heinich, Molenda, Russel, &

Smaldino (1996:37) regard as evaluation being governed by pre-assessment

measures. These pre-assessment measures should be communicated with all

the evaluation participants prior to the evaluation activity.

Ideally this evaluation type should therefore start with a meeting where an

evaluator’s programme description is worked out. Masokoane (1993:64 in

Kromberg, Govender, Birrell & Sibanyoni) refers to this consultative process as a

‘moral and ethical framework of good practice’ and ‘democracy, quality and

transparency’. This description has to match the standards for evaluation. To

measure success, a comparison is made between what is being accomplished

and what was anticipated. This comparison precedes a cost-analysis process in

which the project costs are compared with the cost of similar projects.

2.4.2 Goal-Free Model

As stated by Propham (1974 in Boulmetics and Dutwin 2000:73), the Goal-Free

Model was developed by Scriven who during evaluation looked at a

programme’s actual effect on identified needs. Evaluation is done on what the

project is doing to address the needs of a client or client population. In this

model observation is conducted without the use of a checklist. However, data

is accurately recorded to determine the significance of the findings. At the end

of the evaluation process, the evaluator will have a descriptive analysis of the

results and involve others to participate in such analysis.

The main characteristic of this model is that the evaluator does not commence

the evaluation process with any preconceived notions of the outcome. Scriven

(in Borich 1974:30) describes goal-free evaluation (GFE) as the one in which the

evaluator performs the function from a more general perspective. This also

implies that the organization is not supposed to make any goal statement as this

24

Page 25: SECTION ONE EXPLORATION OF THE STUDY AREA …

might derail the evaluator and obstruct him/her from performing the evaluation

successfully. In this model, the evaluator may employ both obtrusive and

unobtrusive methods to gather data. The former refers to using methods or

equipment that the subjects are aware of (for instance tests), while the latter

refers to those that the subjects are not aware of (for instance hidden tape

recorder).

Based on the above, Rosi and Freeman (in Boulmetics & Dutwin 2000:74),

state that people’s views on the use of the model are different. Some think that

the evaluator has to be an expert in the field to be evaluated whilst others think

such is not necessary. In evaluating LTSM, it is important that the evaluator

possesses skills in evaluation and knowledge of the Learning Area for which

materials are to be evaluated. The lack of the skills and knowledge might

jeopardize the integrity of the evaluation exercise.

2.4.3 Transaction Model

The Transaction Model was developed by Stake, (Boulmetics & Dutwin 2000:75)

concentrating on activity between the evaluator, participant and the project staff

(participants). It is aimed at benefiting the clients and practitioners. In this model

there is a continuous two-way communication (feedback) between the evaluator

and organizational staff. The approach is participative in nature and the

evaluators and clients are viewed as equal partners. The evaluator uses a

variety of interview and observational techniques to gather data. The orientation

can have a goal-free or goal-based technique.

2.4.4 Decision-Making Model

25

Page 26: SECTION ONE EXPLORATION OF THE STUDY AREA …

This model, developed by Daniel Stufflebeam (Boulmetics & Dutwin 2000:75),

is used by leaders to predict the future of the project, which in turn affects the

future of the organization. The concern is about the long range effect of the

project on the organization or the client population. It is not about the present

situation. It assists leaders in making decisions in the future, such as cutting

the costs. Most importantly, the relevant stakeholders are invited to partake in

the decision-making processes.

2.4.5. Goal-Based Model

Due to the fact that it is easy to use, the Goal-based Model is the most

commonly used. Unlike in the goal-free model, in this model the organization is

free to set and present goals or objectives to the project leaders. The goals are

set according to the needs of the organization and plotted on either a proposal,

brochure or project description. Furthermore, Boulmetics & Dutwin (2000:77)

emphasize that the wording of the objectives would usually set and identify the

standards for evaluation. The evaluator uses either qualitative or quantitative

methods to measure the outcomes, objectives or goals that were set by the

client or organization.

Although the above models are interdependent, this research project adopts

mainly the transaction model. This is what the GDE may be aiming at and often

there is mention of social aspects like collaboration and participative ness.

2.5 CRITERIA FOR EVALUATING LEARNING AND TEACHING SUPPORTMATERIALS

Criteria for evaluating LTSM form the basis for the instrument used for evaluation

purposes. However, various authors and experts on evaluation give various

arguments on what needs to be part of the instrument for LTSM evaluation. This

section of the research explores some of those suggested criteria. Most of the

suggested criteria are generic in nature as compared to what is contained in the

26

Page 27: SECTION ONE EXPLORATION OF THE STUDY AREA …

instrument. For this exercise, the suggested criteria will be juxtaposed with

criteria that are contained in the GDE evaluation instrument (GDE LOETA policy

2003:1).

Heinich, Molenda, Russel & Smaldino (1996:36) state that an instrument for

evaluation should allow for a superficial analysis of learner’s characteristics. This

implies that there should be consideration of the learner’s level of development,

age, cultural background and values. Furthermore, there is an argument that if

an instrument makes such consideration this could provide helpful leads in

selecting LTSM.

In the GDE instrument, this coverage extends to almost all the sections. For

instance, under compliance with the curriculum, there is a criterion that deals

with establishing whether the materials are in line with the constitution of the

country in terms of sensitivity to issues of gender, age disability and HIV/ AIDS.

Previously, with some exceptions, the general trend in textbooks and readers,

was over-representation of males on covers, main characters, other characters

and images inside a book (McKinney 2005:13).

Furthermore, McKinney (2005:19) stipulates that most books show a severe

under-representation of black characters and images of disabled persons. The

issue of gender characteristics, particularly masculine and feminine stereotypes

remains an area of concern and needs to be investigated by means of LTSM

evaluation instruments (Evans & Davies 2000: 5).

The reason why the GDE instrument stresses this criterion is probably to avoid

the reoccurrences of the past discriminations. The evaluator is also requested to

consider whether the LTSM cater for learners with varied abilities as stipulated in

the Inclusion/ Education White Paper (GDE 2003: 7).

Furthermore, Heinich et al. (1996:37) argue that in case of a heterogeneous

group, evaluators should be assisted to select LTSM that provide for a common

27

Page 28: SECTION ONE EXPLORATION OF THE STUDY AREA …

experiential base. This is in line with what the GDE instrument illustrates as

context, especially the criterion that seeks to ascertain that learners are given the

opportunity to discover the world that is beyond their own. This connects with

recognition of various social structures; balance between the learners’ familiar

and new information; and acknowledgement of the learners’ context and

experience.

The above mentioned criteria regarding the level of development also include the

language level. This is evident in the argument by Potenza, (in Kromberg,

Govender, Birrell & Sibanyoni 1993:50) that sensitivity to language should be

included be considered when evaluating LTSM. However, oversimplification of

language like English for second language users should be avoided.

In addition to the above criteria, Heinich et al. (1996:45) state that the LTSM

should match the intended outcomes or objectives. At this stage, one is

reminded that the intention is to obtain the appropriate LTSM by selecting from

available LTSM, modifying existing or ‘off the shelf’ LTSM or designing new

LTSM. Irrespective of the category, the LTSM must meet the required standards

in order to be selected.

The above-mentioned criteria would be incomplete without the following checklist

as suggested by Heinich et al. (1996:47):

• Does it match the curriculum?

• Is it accurate and current?

• Does it contain clear and concise language?

• Will it arouse motivation?

• Does it provide for learner participation?

• Is it of good technical quality?

• Is there evidence of its effectiveness (e.g. field-test results)?

• Is it free from objectionable bias and advertising?

• Is a user guide or other documentation included?

28

Page 29: SECTION ONE EXPLORATION OF THE STUDY AREA …

After having gone through the evaluation criteria, one needs to learn that their

application is a specialized process. As implied by Heinich et al. (1996:46),

selection criteria vary with different media formats. In other words, criteria that

apply in one type of LTSM may not necessarily apply in another. For instance,

criteria for selecting a chart may not be the same and not even as detailed as

criteria for evaluating a book or atlas for that matter. To address this need, the

GDE evaluation instrument consists of different sections, each stating the type of

LTSM it may be used for, thus the following:

• Section A: To evaluate Learners’ Books, Learners’ Workbooks/ Activity

Books, Readers, Educators’ Guides and Books

• Section B: To evaluate Learners’ Books, Learners’ Workbooks/ Activity

Books, Readers, Educators’ Guides and Books

• Section C: To evaluate Games, Educational Toys, Models, Puzzles and

Audiovisual Materials

• Section D: To evaluate Charts and Posters

On the other hand, criteria for evaluation may apply to selected, specific learning

programmes, for example the GDE instrument is learning programme specific.

For instance Section B is designed to specifically evaluate LTSM for Literacy,

Life Skills or Numeracy (whichever is applicable).

Some of the criteria suggested by Heinich et al. (1996:46), consider objectives,

audience, cost, technical expertise, equipment and time. It is also suggested

that a preview or screening of LTSM happens before the actual evaluation

process. This could be done by using techniques such as reviews, blurbs or

appraisals by colleagues. This implies that evaluations are done before, during

and after utilization (Potenza in Kromberg et al., 1993:48).

.

29

Page 30: SECTION ONE EXPLORATION OF THE STUDY AREA …

If an education department is able to apply its evaluation criteria, techniques and

skills accordingly, it will be able to eliminate practices such the ones that Potenza

(in Kromberg et al., 1993:48) is objecting to i.e. publishers reprint new editions

of old textbooks instead of publishing new materials. This act could be ascribed

to corruption in the publishing world particularly because some publishers

knowingly submit materials that contravene the laid down standards and criteria

for selection. They often hope to use their unacceptable influences to ensure

that their inappropriate LTSM are selected (Proctor and Monteith in Kromberg et

al., 1993:33).

2.6. SUMMARY

In this section, using different sources, I explored different definitions of the

process of evaluation and constructed my own definition of the concept. I further

tried to contextualize the concept to the evaluation of Learning and Teaching

Support Materials using an instrument designed by the GDE. I also illustrated

that evaluation happens for a specific reason that will benefit the organization or

client population in one way or the other. Some of the prominent reasons for

evaluation are for the organization to increase efficiency, to make sure that

projects are effectively implemented and to increase impact of activities involved

in the project.

In order to achieve the above-named aims of the project, project leaders have to

understand different models and adopt one for their project. However, my

argument is that these models are not exclusive in nature but can be integrated.

The discussion would have been incomplete without an examination of criteria

for evaluating LTSM. To attain that, I explored, compared and contrasted criteria

as suggested by different authors and against the instrument that is used by

GDE to evaluate and select LTSM.

Furthermore, I explored different models for evaluation of LTSM. The implication

is that when one evaluates an LTSM evaluation project, components of each

30

Page 31: SECTION ONE EXPLORATION OF THE STUDY AREA …

model should be considered. A lesson is learnt from Provus’ Discrepancy model

in which there is emphasis on always remembering that a project is not in a

vacuum but part of the organization. Therefore, the GDE LTSM team should

ensure that their project is within certain specific organizational parameters.

This is also in consideration of the fact that a project should be managed in a

participative manner.

Furthermore, a project would be worthless if the needs of the client are not

considered. This is expressesed in Propham’s Goal-free model. The two-way

communication is supported by Stake’s Transaction model in which the

practitioners’ significance is highlighted. This presents all the participants with an

interactive, communicative and collegial relationship.

In this section, there has also been an emphasis on the long-range effect of the

project. Even when the project work has come to an end, the clients or target

population group should continue to enjoy the benefits of the project. This is

expressed in Stufflebeam’s Decision-making model. It is supported by the Goal-

Based model that stresses that the goals should be specific and also meet the

needs of the population.

The next section focuses on the research methodology employed in the study. It

describes the research design and explains the data collection tools which will be

used.

SECTION THREE

RESEARCH METHODOLOGY

31

Page 32: SECTION ONE EXPLORATION OF THE STUDY AREA …

In this section, I will firstly discuss the characteristics of qualitative research in

more detail. Furthermore, qualitative and quantitative research is juxtaposed.

Data collection methods that will be used are also introduced in this section.

Later on in the section, the research design is explained in conjunction with the

population and sample that was used to gather data. There is also an exposition

of the way in which the focus group interviews are conducted. The intention is to

utilize focus groups to promote discourse among the participants. Discourse is

the way to achieve the intended outcome of focus groups which is to share

experience around areas of practice (Wiesener and Mezirow, in Mezirow 2000:

332). This is coupled with the list of questions in the interview schedule. Since a

survey questionnaire is used, it is also vital to discuss its validity and reliability.

3.1. QUALITATIVE RESEARCH

The qualitative approach adopted in this study is characterised by the human

mind that plays the role of being an instrument of research and that is referred to

as positivist discourse of social enquiry (Henning et al.,, Van Rensburg and Smit,

2004:7). This implies that the researcher assumes an Interpretivist role (Trauth

in Henning et al., 2004:9) and that is instrumental in ensuring that the possible

error associated with observation is avoided. The findings of the study are

backed by Literature reviews which evaluate and analyse materials that have

dealt with related subjects in the past.

The aforementioned implies that as opposed to quantitative research, in this

study the participants are free to communicate other information that is not

explicitly expressed on the artefacts (Henning, et al., 2004:3). The variables are

not controlled and the participants are free to respond and naturally develop as

the research work progresses. This is made possible by means of focus group

interviews and observations. Through these data collection methods, the

research is able to delve into depth of the variables.

32

Page 33: SECTION ONE EXPLORATION OF THE STUDY AREA …

The data is gathered by participatory observation, artefact (using survey

questionnaires to evaluate the GDE’s Evaluation instrument) and document

studies and in-depth interviews. Design types include discourse analysis,

interviews and surveys. The design methods are categorised as Ethnographic

studies, Discourse analysis and Qualitative evaluation tool. The researcher is

required to ensure that there is validity, reliability and generalisability in the

instruments and processes that take place (Henning, et al., 2004).

3.2 RESEARCH DESIGN

Having recognized and concluded that teachers who evaluate materials need to

be provided with a useful, valid and reliable evaluation instrument, a research

outline was developed to investigate the matter. The inquiry was initiated with

Foundation Phase teachers who were expected to have interacted with the

evaluation instrument as individuals or groups at different levels including school

level and at the exhibition centre. The logic of the data was that it would make

clear how useful the evaluators find the artefact, i.e. the evaluation instrument.

Since it is expected of schools and teachers in different areas of Gauteng to

have used the evaluation instrument, I decided to gather the data from a

heterogeneous group of 157 potential LTSM evaluators at a training centre

organized by the Ekurhuleni West District of the GDE. This implied that the

inquiry was done on educators from schools that are rated differently as per the

poverty index and were also from different cultural groups.

The selection criteria for evaluator participants were within the following

parameters: teachers currently teaching in the Foundation Phase and who are

involved in teaching based on the Revised National Curriculum Statements

(RNCS). The search for the educators who could be part of the research started

with an interaction with my colleagues in the GDE’s Teacher Development unit.

This interaction involved arranging to have a time slot during the provincial

training of educators on the RNCS to hand out the survey questionnaire and

33

Page 34: SECTION ONE EXPLORATION OF THE STUDY AREA …

explain what was expected of the teachers (e.g. the return of completed

questionnaire) (refer to Addendum D)

To arrange for the above stated, a formal request had to be submitted to the

Senior Manager of Human Resources Development for approval with the aid of

the attached letter to request permission. Conditions for the data collection

process to take place with minimal or no interference with the training process

were negotiated with the teacher development unit personnel. These

negotiations were made possible by accessibility of the GDE Project

Management Team.

Since the permission was granted I went to the training centre and distributed the

artefacts and requested that the training coordinators collect at the end of the

session. This assisted in ensuring that all the completed instruments were

returned. The teachers did not take the instruments home and that made it

possible for a hundred percent collection. Furthermore, I selected a group of

ten participants for a focus interview and also observed them as they trialed the

evaluation instrument against samples of LTSM.

3.3 RESEARCH ORIENTATION AND CONTEXT

The research survey was conducted at the Elspark Secondary School which was

used by the Ekurhuleni East District as a training centre for educators on RNCS

(including how to evaluate LTSM). The educators were requested to complete

and re-submit a survey questionnaire (See Addendum E) on the usefulness of

the LTSM evaluation instrument. Arrangements were made for focus interviews

to be held at a later stage at the different school locations and clusters. The

conditions under which the focus group interviews took place were strictly private

in a room at a school. Interview schedules were designed and used as guidance

during the interviews. This was made possible by the GDE Head Office Teacher

Development unit, the training Project Management Team, Training co-

34

Page 35: SECTION ONE EXPLORATION OF THE STUDY AREA …

coordinators at the training centre and Higher Education Institution

representatives.

3.4 SAMPLING

Although the study is predominantly qualitative in nature, the type of questions

asked in the survey questionnaire is a combination of qualitative and quantitative

inquiry. This questionnaire is administered on the participants who were

identified and formed a sample. These participants were selected by means of a

combination of purposive and theoretical sampling as people who fit the criteria

of desirable participants were selected. These 157 trainees of Ekurhuleni East

District were selected on those premises as they probably possessed the

potential of actively participating in the process in a social interactive way (Refer

to Addendum B).

Furthermore, a smaller group of ten participants was selected to participate in

the focus group interview. During this interview the evaluators were engaged in

a reciprocal social interaction in which they and the researcher shared their pre-

existing knowledge and experience in order to socially construct knowledge on

issues that are pertinent to the evaluation of LTSM in the GDE. It is through this

discourse that the participants learnt that they were equal partners and were also

free to make their contribution in this dynamic social context in which

construction of knowledge took place (Gravett 2001:20).

3.5 DATA COLLECTION METHODS

The primary data collection method was the survey questionnaire and the

secondary data gathering method was the focus group interviews with ten

participants that were randomly selected. Participatory observations were done

as the participants were given the instrument and samples of Foundation Phase

books to evaluate. Participatory observation was used as an experiential frame

during the interpretation of data. This was coupled with what Henning et al.

35

Page 36: SECTION ONE EXPLORATION OF THE STUDY AREA …

(2004:7) term document analysis. A brief discussion of each of these methods

will follow in the passage below.

3.5.1 Survey Questionnaire

As stated above, the primary data collection method is a survey questionnaire

and it was administered prior to the focus group interviews and observations.

Considering the vast nature of the LTSM process as depicted on the process

map above, caution had to be exercised in the designing of the artefact, the

survey questionnaire. The items of focus were first determined prior to the

setting and classification of questions (Henning et al.,, Van Rensburg and Smit

2004:89). This has prompted the classification of the questions into sections,

namely Personal Information, Evaluation Experience and Actual Engagement

with the Instrument.

What is important to note is that the instructions to the participant are on the

initial part of the questionnaire (See Annexure A). In constructing these

instructions, I considered that there was unequal power between myself, the

researcher and the evaluators as participants. Because of my position in the

Department of Education. The evaluators probably regarded me as an authority

and may have been threatened and tense. The instructions were a way of

relaxing them, particularly as they were alerted that they were going to be treated

as partners and would also remain anonymous. The primary aim was to make

process remain a social interaction. That would also ensure that there is true

knowledge sharing (Henning, et al., 2004:67).

3.5.2 Focus Group Interviews

36

Page 37: SECTION ONE EXPLORATION OF THE STUDY AREA …

As stipulated above, the aim of this inquiry is to put together evaluators who have

the necessary pre-existing knowledge and experience to partake in a social

interaction in which new meaning and knowledge is constructed on issues

involving LTSM evaluation and selection. The focus interview therefore creates

a context in which there is reflec tive discussion in which the various participants

bring a range of opinions and experiences as outlined by Morgan (in Fortuin

2002:53).

The questions I asked will be discussed below:

3.5.2.1 Interview schedule

The open ended questions that were asked were as follows:

• Tell me about your understanding of the process of evaluation of the

salient features or steps in the LTSM process map.

• What are the challenges that face the GDE and its institutions in ensuring

that appropriate LTSM are selected?

• If you were given the opportunity what amendments would you introduce

in the LTSM evaluation process?

• Do you regard the involvement of educators in the evaluation of LTSM as

a vital step that was taken by the GDE?

• Tell me about an experience that you will never forget in your involvement

in LTSM evaluation.

• How does the LTSM evaluation instrument offer an educator support in

selecting appropriate LTSM?

3.6 VALIDITY AND RELIABILITY

3.6.1 Validity

37

Page 38: SECTION ONE EXPLORATION OF THE STUDY AREA …

As asserted by Silverman in Fortuin (2002:61), this section is about validity which

is concerned with the accuracy of findings in terms of the reality it reflects. The

GDE evaluation instrument was measured to ensure that it complies with validity

requirements. The validity of this qualitative study is established by the detailed

audit trail which includes evidence of data collection analysis and findings. The

findings as discussed in section three above may be classified as primary data

that was derived from the survey and focus group interviews. The evaluators’

experiences were captured by means of the various data-collection techniques

which were set to possess the validity requirements.

Considering the contribution made by Silverman in Fortuin (2002:61) that

interviews alone can become unreliable at the level of method, I also considered

to precede their utilization with the use of a survey questionnaire which also has

to be a valid data collection instrument. Among other things, I endeavored to

use clear, concise language in both the data collection techniques. The

questions used in the focus group interviews were refined to be as friendly

enough to the interviews. I was quite flexible with the language of

communication as the participants were competent in different languages

although all are comfortable users of the English language. I also applied code-

switching when I gave them some instructions or unpacked the contents of the

survey questionnaire. The transcription of the focus interviews was checked by

my English-speaking colleagues.

3.6.2 Reliability

Reliability may be interpreted as the consistency between the findings and the

methods of data collection and analysis (Merriam, in Fortuin 2002:62). The data

collection methods that were used are appropriate to the research question. The

current study therefore should have some relation with subsequent enquiries in

the same field. All processed data has been made available for further perusal.

38

Page 39: SECTION ONE EXPLORATION OF THE STUDY AREA …

As suggested above, reliability is applied in order to ensure that there is

consistency with the data from focus groups. This assisted in revealing the

experiences of the evaluators of LTSM even when the survey was repeated the

same results were achieved. This also applies to the first and second sets of

focus interviews.

3.7 ETHICS

This study complies with the ethics of the University of Johannesburg as

stipulated in its Academic Ethics Committee Document that was used for

discussion on 5 April 2005. When the survey questionnaire was handed to the

participants, they were told that their participation was voluntary. This was

important because most of them knew the researcher and could have felt obliged

to participate because of his position in the GDE. In addition, the participants

were told about the purpose of the survey: that it was not part of the training

session or a duty to be compulsorily done. Their confidentiality and anonymity

was assured on the first page of the survey questionnaire. All these principles

were adopted for the participants to accept and have trust in, particularly by

ensuring that they do not suspect that they are being betrayed. Furthermore, as

stated in the ethics guidelines of the University of Johannesburg (2005: 1) it was

indicated to the participants that permission to conduct the study was granted by

the Teacher Development Unit.

3.8 SUMMARY

This section was focused on data collection and analysis. The focus of the study

is determined and declared by means of research orientation. The context of the

study is analyzed and stipulated to show the extent of the research and the size

of the population and sample. This is coupled with the explanation of data

collection methods applied, surveys and focus group interviews (using interview

schedules). The concepts of validity and reliability are also discussed.

Furthermore, the interview questionnaire and other instruments are tested for

39

Page 40: SECTION ONE EXPLORATION OF THE STUDY AREA …

effectiveness and efficiency. In the next section data are analyzed and the

research findings emerge.

SECTION FOUR

DATA ANALYSIS AND RESEARCH FINDINGS

40

Page 41: SECTION ONE EXPLORATION OF THE STUDY AREA …

As indicated in the previous section, this inquiry sought evidence for the

usefulness of the instrument used by educators as evaluators of LTSM in the

GDE schools. The previous section focused on the research design and

research methods, namely the research survey and the focus group interviews

and their related tools (survey questionnaires and interview schedules) (Henning

et al., 2003: 7).

In this section, however, research findings will be presented based on the

research literature that was explored in earlier sections. The analysis of data will

be backed by, among others, the work of Henning et al. (2004) in which the

researcher is advised and given skills how to presented data in a narrative way.

There is also an adaptation of the constant comparative method as suggested by

Maykut and Morehouse (1994) in which data is grouped into sub-themes and

themes called codes and categories respectively. Furthermore, there is

implementation of the constructivist approach as viewed by, among others Scott

in Gravette (2001).

4.1 ANALYSIS OF DATA

The findings of this inquiry are a result of the steps undergone in the process.

This process began with a survey that was made possible by means of a survey

questionnaire. The questionnaire is used to develop an understanding of the

challenges experienced by educators as evaluators of LTSM when engaging with

the instrument used to evaluate LTSM. This was followed by focus group

interviews in which ten participants were interviewed for a further data gathering

process. As suggested by Henning et al. (2004: 101), in this section. I shall in

this section make a narrative presentation or analysis of my understanding of the

data that was collected by means of the aforementioned methods.

Interestingly, there is an element of interrelationship between the methods of

data collection, as these themes are covered by the various sections of the

survey questionnaire.

41

Page 42: SECTION ONE EXPLORATION OF THE STUDY AREA …

As indicated below and as suggested by Maykut and Morehouse (1994: 127),

based on the focus group interviews, the following themes were identified. Each

theme develops from codes that were constructed by grouping the responses

given by users of the instrument to evaluate LTSM (Refer Table 1). As the

discussion progresses, it is also important to note how the codes overlap and

thus a code may develop to more than one theme:

• Willingness among educators to perform the function as part of theirwork

This theme has some relationship with and develops from the codes identified

as:

Inadequate support to institutions, Instrument used by different sections in the

GDE, Instrument well marketed within the GDE, Effective training sessions,

Instrument addresses the needs (including curriculum), Instrument enables the

educator to decide on LTSM, Instrument used in collaboration, Instrument boosts

confidence in evaluation and Evaluation is indispensable.

• Evaluation happens with some assistance

The corresponding codes for this theme are: Inadequate support to institutions,

Effective training sessions, Instrument used in collaboration, Instrument boosts

confidence in evaluation, and Evaluation is indispensable.

• Instrument is the core element for evaluation process

This theme relates with codes like: Instrument used by different sections in the

GDE, Instrument well marketed within the GDE, Effective training sessions,

Instrument addresses the needs (including curriculum), Instrument enables the

educator to decide on LTSM and Instrument boosts confidence in evaluation.

42

Page 43: SECTION ONE EXPLORATION OF THE STUDY AREA …

• Instrument encourages collaborative evaluation

This section makes provision for statements like: Instrument is used by different

sections in the GDE; Instrument is well marketed within the GDE; Instrument

contributes to effective training sessions; Instrument addresses the needs

(including curriculum); Instrument enables the educator to decide on LTSM;

Instrument is used in collaborative processes; and Instrument boosts confidence

in evaluation, and Evaluation is indispensable.

As illustrated by Maykut and Morehouse (1994; 127), the constant comparative

method is a useful method to analyse data. In this form of analysis data is

interpreted and categorised for the purpose of illuminating and elucidating the

research problem that would have been stated by the researcher. The data is

further sorted and named, with a constant comparison of new codes and

categories (Fortuin (2002; 56)

In relation to this study, the above codes were identified and analysed

considering the social constructivism basis of this argument. The manner in

which the evaluators of materials engage with the instrument and the types of

responses that form codes and categories in the above table, are evaluated

against the premises of social constructivism. The evaluators’ prior knowledge

of evaluation is accepted and utilized as a useful aspect of the evaluation

process. To make possible a social interaction, the project leaders do not use an

individualistic approach in which evaluators are required to work as individuals.

Instead, the evaluators are put in a room in which they are able to interact with

one another. During that interaction, the evaluators learn from one another and

together create meaning and knowledge of what is expected of them as

evaluators. This also applies to the way in which they put together reports on the

LTSM that they are evaluating.

When the focus group interviews take place, the aforementioned atmosphere

that promotes interaction makes it easy for the evaluator to codify the

43

Page 44: SECTION ONE EXPLORATION OF THE STUDY AREA …

participants’ responses to some of the questions which are based on the survey

questionnaire. The researcher makes the process as participative as possible by

relaxing the evaluators so that they are not conscious that he is recording the

interviews. At the end of each session, their opinions are also considered.

In my attempts to explain, I had to systematically proceed from the organizing

and coding of data, grouping of data into categories, to developing rules in order

to include all the units in each category and lastly reflect on the responses to the

research question (Fortuin 2002:56).

Reading through the focus interview transcriptions (see Addendum F) enabled

me to have a better understanding of the frustrations and problems experienced

by evaluators of materials in their engagement with the evaluation instrument.

Their responses included expressions like:

• “need more support”

• “better when working with others”

• “feel more secure”

• “managers hide away information”

• “learn a lot from Evaluation Centre”

• “need to do it more often”

• “once per annum is not enough”

• “no sense when education department does it for us”

• “feel owning the process”

• “educators should receive even more incentives”

• “should be part of our careers”

These responses were classified and categorised into the codes indicated in the

above discussion.

4.2. DISCUSSION OF THE FINDINGS

4.2.1 Motivation for educators to be evaluators

44

Page 45: SECTION ONE EXPLORATION OF THE STUDY AREA …

It is interesting to note that during the focus group interviews, some evaluators

expressed a wish for the evaluation function to form part of their work and be

included in their job descriptions. However there seemed to be mixed feelings

on this issue. For instance, Simon1, a respondent stated that: “remember aseducators we receive meagre funds… I am therefore not willing to work forno pay” (see Addendum F).

The above statement implies that some educators view themselves as a group

distinct from their employer. This also implies that among educators there is still

a “we-they” polarity as portrayed by Burtman and Borich (Borich 1974:39).

However, such educators do not consider the fact that evaluation is a means to

empower them in the name of transparency and democracy as suggested by

Masokane (1993:64 in Kromberg, et al.). It can also be concluded that some

educators take the formal evaluation process as an opportunity to raise some

income. This notion is supported by the view of Borich (1974:27) that evaluation,

like the cost of living, has as its aim a propensity for rapid change.

Nevertheless, I argue that we may not conclude from this statement that among

the educators there is total unwillingness to partake in the evaluation process.

This is because the respondent does not complain about being part of the

process. Evidence for that willingness is strengthened by Sandile’s remark when

he said: “remember that it is your calling. Besides, it is within the principlesof democracy that we should have a contribution to make in ourprofession” (see Addendum F). This statement also implies that some

educators would like to have the evaluation process as a non-optional

professional demand by the GDE on its educators (Kellaghan, et al., 2003:17).

Besides for the above-stated reasons, some educators are eager to have

appropriate materials in their classrooms. They think by forming part of the

evaluation teams they will own the process and will, as Mary said: “…selectmaterials that do not have any weaknesses that could hamper or negatively

1 In this study I will use pseudonyms in order to preserve the anonymity of the respondents

45

Page 46: SECTION ONE EXPLORATION OF THE STUDY AREA …

affect the teaching and learning processes in our schools” (see AddendumF). Besides, this involvement will also entitle the educators to a role of

partnership with education officials.

Furthermore, educators as participatory evaluators of materials claim that even if

they may engage in voluntary evaluation or perform that as part of their work,

they will still need the backing of senior officials in the department. Sandile

voiced that when he said: “… we will feel more secure and protected if wehave the backing and support of our line functionaries in particular if ourprincipals could refrain from blocking dissemination of information” (seeAddendum F). In other words this statement shows a concern on outcomes

which reveal the impact of their involvement and participation in the process of

evaluation. This is in line with the view that impact refers to the attainment and

non-attainment of outcomes (Borich 1974: 88).

As implied earlier on evaluators are engaged in different kinds of discourse.

First, it is their discourse with the text or titles that they evaluate. Second, it is

the discourse with fellow evaluators of materials. The internal evaluators are

willing to have such a discourse on a daily basis. At school level, educators who

are evaluators can enter into contrastive discourse which could provide socially

relevant findings and insights which could be instrumental in their selection of the

LTSM on behalf of the school (House 2000:598).

As suggested in section two above, the aforementioned still implies that

educators are involved in a dialogic construction of knowledge with their fellow

colleagues (Gravett 2001:20). This is simply because the instrument

encourages an educator to engage with others in the process of evaluating LSM.

This kind of discourse manifests itself in the form of discussions.

The notion that other respondents argue that the evaluation process is a

professional demand by the GDE on its educators, was evident when Maxwell

said: “At times I think they are abdicating their responsibilities because

46

Page 47: SECTION ONE EXPLORATION OF THE STUDY AREA …

they leave us without support at all” and in support, Sandile stated: “it iswithin the principles of democracy that we should have a contribution tomake in our profession. We need to be proud by knowing that we werepart of the selection process” (see Addendum F). The general feeling is that

the evaluators are bound by the predetermined standards decided by the

curriculum and reflected on the instrument for evaluation. This also correlates

with what was stated earlier that for effectiveness, efficiency and impact, the

process should not be goal - free as per the description given by Borich

(1974:30). This finding links with the following one as the evaluators feel that

while they are willing to perform the function they need some support in the form

of collaborative interaction.

4.2.2 Supported evaluation

According to the social constructivist viewpoint adult learners should be given an

opportunity to construct new knowledge utilizing their pre-existing knowledge and

experience by recognizing and reconstructing experiences encountered on the

social plane (Driver-Scott in Gravett 2001:20). Further, they use these

experiences to enrich and revise their knowledge. The need for this kind of

support is expressed by the evaluators of LTSM. They expect the GDE to create

such a social plane that comes in the form of a partnership between the LTSM

coordinators and themselves. This is evident from Bafana’s response when he

said:

Based on my experience as an educator, member of the materialscommittee and evaluator of materials, I would think that the GDE ismoving towards the right direction. In the past we were excludedfrom the process of taking decisions as to what kind of materials wewished to utilize to develop our children. Materials that we didn’tdesire were ordered on our behalf. We were not given theopportunity to accept or reject (see Addendum F)

47

Page 48: SECTION ONE EXPLORATION OF THE STUDY AREA …

This statement remains constructive for as long as evaluators do not enter the

process with preconceptions (Scriven in Borich 1974:30).

The evaluators do acknowledge the fact that their pre-existing knowledge and

experience are pivotal in the process of evaluating LTSM. However, even

experienced evaluators are not confident enough without some kind of support

that they expect the GDE officials to render. This is evident from Jeremy’s

response that: “It is actually a very complicated process when you are notgiven support”. Furthermore, Maxwell said: “Especially the so called districtofficials. At times I think they are abdicating their responsibilities becausethey leave us without support at all” (see Addendum F). I could argue that

this statement indicates a certain amount of insecurity on the part of some

evaluators. Jeremy’s point is actually refuted by the view that evaluators should

make their contributions that are influenced by their engagement in the

evaluation process as stated by Stufflebeam (in Kellaghan et al., 2003:33).

In fact, during the focus group interview interaction most of the interviewees

expressed a wish to partake in evaluation that takes place at the provincial level

rather than at school level where they believe there is lesser support. This is

clear from Hlomu’s response when he said: “They advise us to do that even atschool level. In this case you need to consult with other teachers via theLTSM committee to decide on the kind of materials to be selected for yourclassroom” (see Addendum F). However this is refuted by one of the premises

of the discrepancy model by Malcolm Provus (in Boulmetics and Dutwin 2000: 6)

which states that a project is not in a vacuum but falls within an organizational

structure.

From the above-stated responses one may conclude that there is a general

feeling that District offices do not render the expected support regarding the

evaluation of LTSM. However, the evaluators are conscious about the fact that

they need to have an input in form of academic and professional skills,

knowledge and experience (Stufflebean in Kellaghan et al., 2003:31).

48

Page 49: SECTION ONE EXPLORATION OF THE STUDY AREA …

The efforts that the evaluators put into the evaluation process would be fruitless

if they did not have reliable and useful artefacts. This forms part of the next

theme of the findings of this inquiry.

4.2.3 The Instrument forms the core element of the evaluation process

During the focus group interview evaluators who were involved in the evaluation

of LTSM prior to the introduction of the GDE evaluation instrument regarded the

instrument as a useful artefact. George, who had some experience, supported

that when he was asked whether he thought the instrument was useful, he said:

Oh! That instrument is my saviour. You know, I have always hadpassion for the exercise of evaluating the materials. My biggestfrustration was doing that without any instrument. I remember wewould sit together as a team and create our own criteria – some ofthose were not measuring what we intended to find out about thebook (see Addendum F).

George, in his statement, could be referring to the effectiveness which is in line

with what was cited earlier from the work of Boulmetics and Dutwin (2000: 25)

which states that an evaluation takes place to determine whether an action was

effective or not. Furthermore, this is an exposition of the need to have pre-

assessment measures that govern evaluation as suggested by Heinrich, et al.

(1996:38).

Absence of an instrument is probably as bad as having one that is subjective.

Jenny’s statement supports George’s earlier assertion when she said: “Iremember that took away a lot of trust. We were in the schools’ textbookcommittee and some staff members felt that we were favouring certaintitles at the expense of others” (see Addendum F). This is also in line with

the notion that in any project there is a set of standards stipulated to examine

49

Page 50: SECTION ONE EXPLORATION OF THE STUDY AREA …

performance (Heinrich, et al., 1996:37). The tendency was for them to express

the frustrations that they experienced in their endeavours to evaluate and select

LTSM for their classrooms. Some evaluators, as stated in the previous section

felt that without an evaluation instrument, the GDE would be denying them some

support.

Considering the qualities of the instrument, responses on the survey

questionnaire indicate some positive attitude towards the evaluation instrument.

This is also expressed on the group findings where 46% of the respondents feel

that a specific section on the instrument should not be changed they are happy

with the organization of the instrument.

The general aspects of the instrument include aspects like the LTSM’s physical

aspects (durability, safety, binding, and texture), colourfulness, font size, white

space, and objectivity. The specific aspects of the instrument test for the

technical, professional and educational qualities (including compliance with the

curriculum, relevance to the Learning Area, relevance to the assessment

requirements and the level of the learners’ development (GDE 2003:7-15).

Objectivity may further be ensured by avoiding the publishers’ knowledge of who

the evaluators are prior to finalization of the evaluation process (Denning 1992:

9).

The evaluators, particularly the less experienced, seem to be happy with both

aspects. However, the more experienced evaluators, some of whom received

formal evaluation training at tertiary level, have expressed some gratitude for the

specific, technical aspects of the instrument. They have even expressed that

this aspect was not explicit enough in the other kinds of instruments or

instructions that they were required to utilize before.

On the survey questionnaire, the 157 participants were asked whether they were

familiar with the GDE instrument. A total of 88 participants (56%) stated that

they were not familiar with the instrument at all while 69 (44%) were familiar with

50

Page 51: SECTION ONE EXPLORATION OF THE STUDY AREA …

it. The participants’ responses were grouped using the tally system, converted to

frequency. Those that were familiar with the instrument used it at varying

frequencies per annum, either once, twice, quarterly or monthly. These varying

frequencies are caused by a number of reasons some of which could be

obtained during focus group interviews. One may conclude that they find the

instrument to be an important element in the evaluation of LTSM.

When the educators were asked if the instrument needed to be improved or

changed, the majority thought that it needed to be adjusted. This is supported by

the findings as discussed in the next paragraph. Fikile, one of the respondents

expressed it more specifically when she said: “Our education system isdynamic and therefore whatever instrument is used needs to be adjusted tokeep up with the times”. Such adjustments are pivotal in ensuring that the

instrument keeps up with the needs of the client or client population as

suggested by Propham (in Boulmetics and Dutwin 2000:73) in his Goal-Free

model.

The findings on this issue were grouped and summarised as follows: of the 157

respondents, eight that the information on the front page needs to be updated as

there are gaps of lacking information, 16 of them thought that the notes that are

made to the evaluator need to be revised and probably made more concise, 41

were of the opinion that the section on the evaluation of readers, games and

charts should be maintained whereas 17 thought that the section needed to be

changed for better, seven think that the instrument is not long enough and needs

to be lengthened while 16 do not agree with that and think that the instrument is

too long and should be shortened. The remaining 52 respondents opted not to

express any opinion in this category of questions.

In conjunction with the earlier section on literature review, the aforementioned

reflect on the level of the instruments’ effectiveness, efficiency and impact. In

particular, it is to ascertain whether the instrument does enable the evaluators to

51

Page 52: SECTION ONE EXPLORATION OF THE STUDY AREA …

competently select only appropriate LTSM. That is whether it introduces

substantive changes in educators’ knowledge, skills and attitudes.

Another area of concern that was investigated by this study was whether the

instrument is used and if so, how often. Of the 157 respondents, 28 indicated

that they use the instrument once per year, six use it twice per year, 10 once per

quarter, 10 once per month, 15 more than once per month, 88 either do not use

the instrument at all or are not sure how often they use it.

4.2.4 Fourth theme: The Instrument encourages collaborative evaluation

One of the prominent ideas that were observed in this inquiry is that evaluators

of LTSM find it more sensible and more effective when to work collaboratively

with their colleagues. The model followed and adopted by the GDE is in line with

that situation at different levels where evaluation of LTSM takes place. The

setting for the evaluation process is not a natural one but is arranged and

organised by the evaluation coordinator whether at school, district or provincial

level. As stated in the earlier section on literature review, this kind of setup is in

compliance with what is asserted by Boulmetics and Dutwin (2000:2) that

evaluation happens mostly in controlled environments.

The findings reflect that 20% of evaluators who use the instrument rely on some

assistance from others. Although this does not form a majority of the

respondents, there is evidence that social interaction is the way in which

evaluators prefer to work during evaluation. This is also evident as, Victor one of

the respondents stated that:

When you are with your peers you are able to give yourselvesautonomy, modify the instrument and convert the quantities toqualities. In this situation you are able to discuss with your

52

Page 53: SECTION ONE EXPLORATION OF THE STUDY AREA …

colleagues some issues that would not have been catered for by theinstrument (see Addendum F)

This correlates with what Jeremy from one of the focus groups says that: I thinkwe do a wonderful job when we work in our teams at school and atprovincial Evaluation Centres (see Addendum F). These perceptions are in

line with the view of Stake (in Boulmetics & Dutwin) in his transaction model

where he states that activity between the evaluator, organizational staff and

fellow evaluators is of significance.

The conclusion made above could have some relationship with the following

responses that were collected after the users of the evaluation instrument were

being asked about the complexity and usability of the instrument. In their

responses, 53 respondents thought that the instrument was easy to follow, 36

felt that it was difficult to use, 14 indicated that they do follow the way it should

be used provided they are given assistance, 6 thought that the instrument was

not usable at all, 47 were not sure or do not use it at all, and only one respondent

did not respond.

As illustrated in Section One of this document, at the Evaluation Centre

evaluators are divided into teams of three. This implies that every title will be

evaluated by three people. Each team comes together and discusses all the

relevant issues and sets up standards that will be considered in this process.

Each team member is therefore given a copy of the title and is required to spend

quality time reading the LTSM and using the instrument to perform evaluation.

Later on, the team meets and consolidates reports into a single composite

report. This is also in line with the constructivist conviction that meaning-making

activities of an individual do not take place in isolation but in a context that is

shaped by societal influences (Gravett 2001:20). Similarly, Vygotsky, the father

of constructivism, concurs that mental function is initially external (intermental)

and later on internal (intramental) (Wersch & Toma in Gravett 2001:20).

53

Page 54: SECTION ONE EXPLORATION OF THE STUDY AREA …

Secondly, at school level activities happen at Macro, Meso and Micro levels. At

the macro level the school and its community are participant; at the meso level it

is heads of departments and other committee heads; and at the micro level the

Learning Area educator. These school activities include evaluation and selection

of LTSM. An educator who is an evaluator finds him/herself involved in a

collaborative relationship. Evident from the focus group interview, it is clear that

evaluators would find evaluation a difficult and tormenting task if there was no

team work and collaborative construction of knowledge. This kind of encounter

is a collaborative inquiry (Wiessner and Mezirow in Mezirow 2000: 331).

4.3 SUMMARY

In this section, I conducted an analysis of the research findings using a

technique known as constant comparative analysis. Each of the findings was

placed as a category of its own and interrogated against the knowledge found in

literature review. Most importantly I illustrated how the responses obtained in the

focus group interviews contribute to the themes which lead to the conclusions

and recommendations that will be discussed in Section Five below.

SECTION FIVE

CONCLUSION AND RECOMMENDATIONS

54

Page 55: SECTION ONE EXPLORATION OF THE STUDY AREA …

In this inquiry, the aim was to establish the level at which GDE educators as

evaluators engage with the instrument used to evaluate and select LTSM. The

roots of the practice of evaluation and selection are based on the international,

national and local activities around the evaluation and selection of LTSM. In the

earlier section of this inquiry, data collection methods like the survey

questionnaire and focus interviews were used to gather the evaluators’

responses regarding the evaluation and selection process. The findings as

discussed in Section Four above will influence the conclusions and

recommendations to be made in this section.

5.1 CONCLUSION

Based on the findings one should be confident to make the assertion that

through the GDE’s utilization of the Evaluation instrument, educators as

evaluators, at different levels, are able to engage with the instrument and select

appropriate materials for their classrooms to a certain level of success.

However, the evaluation process is more meaningful when the educators engage

in some sort of a team work or collaboration which will also prompt mutual

assistance. For that to happen, the instrument used for evaluation is the core

element or artefact. It is also striking to learn that most educators are willing to

perform the function of evaluating and selecting LTSM as part of their work.

During the focus interviews, it was clear that the educators as evaluators were

gratified for the fact that the whole process of evaluation and selection posed an

opportunity for hem to engage with their pre-existing knowledge and experience.

The pre-existing knowledge might have been gathered when they underwent

training as educators, during In-Service-Training, or at their work stations. In

other words, they were given an opportunity to construct new knowledge utilizing

their pre-existing knowledge and experience by recognizing and reconstructing

experiences encountered on the social plane (Driver Scott in Gravett 2001:20).

The pre-existing knowledge and experience are pivotal in the LTSM evaluation

process. Through this inquiry, I have established that the GDE has, to some

55

Page 56: SECTION ONE EXPLORATION OF THE STUDY AREA …

extent, created a social plane that comes in the form of a partnership between

the LTSM coordinators and themselves. It has also been established that some

educators welcome the support and assistance that is offered by the GDE.

The evaluation and selection process would be impossible to pursue without the

existence of a curriculum-compliant, user friendly instrument. Evaluators of

LTSM find it easy to work with an instrument that is well-organised into sections.

The evaluators who have used the GDE instrument stated that the GDE

evaluation instrument does comply with this aspect and include technical,

professional and educational qualities (including compliance with the curriculum,

relevance to the Learning Area, relevance to the assessment requirements and

the level of the learners’ development (GDE 2003:7-15).

Furthermore, I conclude that the GDE should encourage evaluators to work more

collaboratively with their colleagues when performing evaluation. In the focus

interviews some of them expressed satisfaction with the model that the GDE has

adopted in which they work in teams. This may have increased the educators’

willingness to perform the evaluation process more regularly than they do

presently, even if it forms part of their day to day teaching obligations.

5.2 RECOMMENDATION

As part of the conclusion, it is important to make some recommendations that

may directly and indirectly benefit all the participants in the process of evaluating

LTSM. This includes evaluators, learning area educators, parents, school

governing body representatives, school management teams, GDE district and

head office LTSM units including any other person who is involved in the

evaluation and selection of LTSM.

Considering the level of commitment by educators in this process, I would appeal

to the GDE to be more conscious of the significance of this process. I recognise

the existence of the SLTSMCs which serve as some form of the GDE’s

56

Page 57: SECTION ONE EXPLORATION OF THE STUDY AREA …

recognition of the function. The manner in which passion for the evaluation of

LTSM is expressed by some evaluators prompts me to recommend that the GDE

formalizes the existence of the LTSM department in the schools’ organizational

chart. This would relieve the LTSM Educator or head from other activities that

may derail him/her from the focus of ensuring that quality LTSM are selected for

the schools.

As stated above, educators have a wish to perform the function more regularly.

This interconnects with the above-stated recommendation of formalizing the

LTSM structure at schools. Such an ideal would only possible if the latter is also

achieved. Furthermore, the GDE has introduced Gauteng-On-Line through

which Information and Computer Technology Infrastructure is installed at schools

according to a specific plan. The GDE needs to equip itself by composing

criteria for evaluation and selection of software as it is expected of materials

developers to approach the GDE in their bid to have their electronic LTSM

utilized in the schools’ classrooms. This would also assist the GDE to cope with

the possible development from printed to electronic LTSM. This would also

imply utilization of technologies as suggested by the Eastern Cape Department

of Education (2003: 17).

Lastly, in support of collaboration and in order to be sensitive to the evaluators’

need to work in teams, I would also like to recommend that the GDE, at the

Evaluation Centre, maintains the model of working in teams. If the model is

appropriately utilised, the process of evaluation and selection of LTSM could be

more effectively performed. By and large, one may conclude that the findings of

this inquiry have revealed that evaluators do need to be involved in collaborative

construction of meaning, using the evaluation instrument (which they regard as

useful). If LTSM evaluation is worthwhile, they are prepared to undertake it as

part of their daily duties.

57

Page 58: SECTION ONE EXPLORATION OF THE STUDY AREA …

LIST OF REFERENCES

Boulmetics, J. & Dutwin, Phylis (2000). The ABCs of evaluation – timeless

techniques for progam and project managers . San Fransisco; Jossey-Bass Inc.

58

Page 59: SECTION ONE EXPLORATION OF THE STUDY AREA …

Booth,W., Colomb,G. & Williams, J. (1995). The craft of research. In E.

Henning & S. Gravett (Eds), Perspective on lifelong learning/ Research and

writing composition. Melville:RAU Centre for Distance Education.

Brown, B.L. (1998). Applying constructivism in vocational and career education.

Information series, n. 378. Columbo. Ohio:ERIC clearinghouse on adult, career

and vocational education.

Caffarella, P.S. (1998). Planning programmes for adults. An interactive

process. Adult learning, 10(2).

Cervero, R.M. and Wilson, A.L. (1998). Reflecting on what program planners

do. In Cookson, P.S. (Ed.), Program planning for the training and continuing

education of adults (135-163). Malabar:Krieger Publishing.

Charmaz,K. (2002). Qualitative Interviewing and Grounded Theory Analysis . In

J.F. Gubrium and J.A. Holstein (eds), Handbook of interview research. Context

and Method. (679- 694). London:Sage .

Cookson, P.S. (Ed.) (1998). Program planning for the training and continuing

education of adults. North American Perspectives. Malabar:Krierger Publishing .

Czerniewicz, L. (1993) Learning lessons from African publishing. In Kromberg,

Govender, Birrell and Sibanyoni (Eds.), Publishing for democratic education (67-

80). Cape Town:Sached Trust.

Daloz, L.A.P. (2000). Transformative learning for the common good. In

Mezirow, J. 7 & Associates (Eds.), Learning as transformation:Critical

perspectives on a theory in progress. San Francisco:Jossey-Bass.

Denning, C. (1992). South Africa- Textbook sector study. International Book

Development Ltd: 1992.

59

Page 60: SECTION ONE EXPLORATION OF THE STUDY AREA …

Eastern Cape Department of Education (2004). Review of LSM supply chain).

Bisho: ECDE

Evans, L. & Davies, K. (2000). No sissy boys here: a content analysis of the

representation of masculinity in elementary school reading textbooks. In Sex

roles: A journal of research v42 n3-4 p255-70.

Engerstrom, Y. (1991). Activity theory and individual and social transformation.

Activity Theory 7(8), 6-17.

Fortuin, B. (2002). The learning experiences of first time e-learners. Melville:

RAU Faculty of Education.

Gauteng Department of Education (2003). Foundation Phase Evaluation

Instrument. Johannesburg: Gauteng Department of Education Head Office

Gauteng Department of Education (2002). Learning and Teaching Support

Materials policy. Johannesburg: Gauteng Department of Education Head Office

Gauteng Department of Education (2003). Limited Open Equivalent to Approval

policy. Johannesburg: Gauteng Department of Education Head Office

Gauteng Institute for Education Development – Gauteng Department of

Education (

Gravett, S. (2001). Adult learning:Designing and implementing learning events-

A dialogic approach. Pretoria:Van Schalkwyk.

Gravett, S. (2001). The design and implementation of learning events for

adults- Study Guide. Mellville:RAU Centre for Distance Education.

60

Page 61: SECTION ONE EXPLORATION OF THE STUDY AREA …

Henning, E., Van Rensburg, W.A.J. & Smit, B. (2004). Finding your way in

qualitative Research. Pretoria:Van Schaick Publishers.

Heinich, R, Molenda, M, Russel, .D. & Smaldino, S.E. (1996). Instructional

media and technologies for learning. Ohio:Prentice-Hall.

House, E. & Howe, K.R. (2003). Deliberative Democratic Evaluation. In

Kellagan, Stufflebeam & Wingate (Eds.), International handbook of Educational

Evaluation (79-103). Dordrecht:Kluwer Academic Publishers.

Kvale, S. (1996). Interviews. An Introduction to qualitative research

interviewing . London:Sage.

Lincoln, Y. (2003). Constructivist Knowing, Participatory Ethics and Responsive

Evaluation:A model for the 21st Century. In Kellagan, Stufflebeam & Wingate

(Eds.), International handbook of Educational Evaluation (69-78).

Dordrecht:Kluwer Academic Publishers.

Masokoane, G.U. (1993). Tackling racism in textbooks In Kromberg, Govender,

Birrell and Sibanyoni (Eds.), Publishing for democratic education (61-66). Cape

Town:Sached Trust.

Maykut,P. and Morehouse, R. (1994). Beginning qualitative research: a

philosophical and practical guide. London: Falmer Press

McKinney, C. (2005). Textbooks for diverse learners- A critical analysis of

learning materials used in South African schools. Cape Town: HSRC press.

Meloy, J.M. (2002). Writing the qualitative dissertation – understanding by

doing. New Jersey: Lawrence Earlbaum Associates Inc.

61

Page 62: SECTION ONE EXPLORATION OF THE STUDY AREA …

Menton, M & Hiebert, R.A. (1999). Learning disabilities research and practice.

Utah: University of Utah

Mezirow, J. (2000). Learning to think like an adult: Core concepts of

Transformation theory. In Mezirow, J. & Associates (Eds.), Learning as

transformation:Critical perspectives on a theory in progress. San

Francisco:Jossey-Bass.

Monyokolo,M. (1993). Financing and provision of textbooks in South African

Black Schools. Johannesburg: University of Witwatersrand

Oliphant, A.W. (993). Arguments for a strong, indigenous publishing industry.

In Kromberg, Govender, Birrell & Sibanyoni (Eds.), Publishing for democratic

education (67-106). Cape Town:Sached Trust.

Omolewa, M. & Kellaghan, T. (2003). Educational Evaluation in Africa. In

Kellagan, Stufflebeam & Wingate (Eds.), International handbook of Educational

Evaluation (465-479). Dordrecht:Kluwer Academic Publishers.

Potenza, E. (1993). It doesn’t make sense: educational standards in textbooks.

In Kromberg, Govender, Birrell & Sibanyoni (Eds.), Publishing for democratic

education (47-60). Cape Town:Sached Trust.

Proctor, A. & Monteith, M (1993). Textbooks, corruption and curricula In

Kromberg, Govender, Birrell & Sibanyoni (Eds.), Publishing for democratic

education (31-46). Cape Town:Sached Trust .

Phillips, N. & Hardy, C. (2002). Discourse Analysis. Investigating process of

social construction. Qualitative research methods series 50. London:Sage.

62

Page 63: SECTION ONE EXPLORATION OF THE STUDY AREA …

Scriven, M. (2003). Evaluation theory and Metatheory. In Kellagan,

Stufflebeam & Wingate (Eds.), International handbook of Educational Evaluation

(15-30). Dordrecht:Kluwer Academic Publishers.

Sork, T.J. (2000). Planning educational programmes. In A.L. Wilson & E.

Hayes (Eds.),. Handbook of adult and continuing education. San

Francisco:Jossey-Bass.

Serguin, R. (1989). The elaboration of school textbooks – Methodological

guide. UNESCO.

Shailaja, M. & Hiebert, E. (1999). Literature anthologies: The task for first-

grade readers. Michigan:

Sork, T.J. (1998). Program priorities, purposes, and objectives. In Cookson,

P.S. (Ed.), Program planning for the training and continuing education of adults

(273-299). Malabar:Krieger Publishing.

Stake, R. (2003). Responsive Evaluation. In Kellagan, Stufflebeam & Wingate

(Eds.), International handbook of Educational Evaluation (63-68).

Dordrecht:Kluwer Academic Publishers.

Stufflebeam, D. (2003). The CIPP Model for Evaluation. In Kellagan,

Stufflebeam & Wingate (Eds.), International handbook of Educational Evaluation

(31-62). Dordrecht:Kluwer Academic Publishers.

United Nations Commission for Human Rights (1998). Universal Declaration of

Human Rights:general Assembly Resolution 217(III), Article 26.

University of Johannesburg (2005). Academic ethics committee document for

discussion. Mellville: UJ.

63

Page 64: SECTION ONE EXPLORATION OF THE STUDY AREA …

Wiesener, C.A. & Mezirow, J. (2000). Theory building and the search for

common ground. In Mezirow, J. & Associates (Eds), Learning as

transformation:Critical perspectives on a theory in progress. San

Francisco:Jossey-Bass.

Wright, C. (1993). Racism and sexism:legal considerations. In Kromberg,

Govender, Birrell & Sibanyoni (Eds.), Publishing for democratic education (67-

80). Cape Town:Sached Trust.

ADDENDA

64

Page 65: SECTION ONE EXPLORATION OF THE STUDY AREA …

ADDENDUM A: EVALUATION INSTRUMENT

Evaluation Instrument for Learning Support MaterialsFor Foundation Phase

(REVISED NATIONAL CURRICULUM STATEMENTS)2004Learning area: Grade:

Publisher:Language:

Author/s :

Title: Number of components:

Subtitles: ISBN: Price:

Evaluators:Name Signature

Educators’ Guide Learners’ Book Work Book Reader(s)

Audiovisuals:Charts Poster Models Toys Puzzles Games

65

Page 66: SECTION ONE EXPLORATION OF THE STUDY AREA …

Team Leader(s)Name Signature

EVALUATION TOTAL TICK CROSSGeneric Evaluation- Section ALearning Area Evaluation – Section B

FINAL DECISION: Approve two ticks Approved � Not approved �

Date: ____________________

66

Page 67: SECTION ONE EXPLORATION OF THE STUDY AREA …

SECTION A GENERIC EVALUATION

Rating scale:Very little = 1 Partially = 2 Mostly = 3 Very much so = 4Circle the score that best describes the material/s.

1.Compliance with the RNCS Weight= 10%

1.1 The Critical and Developmental Outcomes are covered in the LTSM.1.2 All the Learning Outcomes of the Learning Area are covered effectively and comprehensively in terms of content and activities.1.3 The materials are in line with the Constitution in terms of sensitivity to issues of poverty, inequality, race, gender, age, disability and HIV/AIDS.1.4 The materials cater for learners with varied abilities, as stipulated in the policy on inclusion / Education White Paper 6.

110

16Score Total ×

1 2 3 4 1 2 3 4

1 2 3 4

1 2 3 4

2.Assessment Weight= 10%

2.1 The assessment activities accommodate diverse contexts, e.g. diverse cultures, environments and religions.2.2 The assessment activities allow for learners to integrate and apply knowledge and skills. 2.3 Different assessment methods have been used, e.g. self, peer, group, educator, etc to cover the various aspects of the learner’s performance.2.4 The Assessment Activities assist the learners to set their own goals for progress. 2.5 The material assess thinking skills at different levels

110

20Score Total ×

1 2 3 4

1 2 3 4

1 2 3 4

1 2 3 41 2 3 4

3. Integration of Learning AreasWeight = 5%

3.1 The principles of the RNCS such as social justice, healthy environment, Human rights etc. are inherent in the LTSM and integrated across various Learning Areas.3.2 The Learning Outcomes and Assessment Standards of different Learning Areas are integrated in the different activities.

15

8Score Total ×

1 2 3 4

1 2 3 4

67

Page 68: SECTION ONE EXPLORATION OF THE STUDY AREA …

Rating scale:Very little = 1 Partially = 2 Mostly = 3 Very much so = 4

Circle the score that best describes the material/s.

5. Progression and Relationship Weight = 5%

5.1 There is progression within and between activities eg. the activity on identifying dangers at home should precede a discussion on how to maintain safety at home.5.2 There is a meaningful conceptual relationship between activities.

15

8Score Total ×

1 2 3 4

1 2 3 4

6. Context Weight = 10%

6.1 The learner’s prior knowledge in terms of experience and context is sufficiently acknowledged as a starting point.6.2 The content and activities provide opportunities for learners to discover a world beyond their own.6.3 There is a balance between information familiar to the learners and new knowledge both in illustrations and in text.6.4 Provision has been made for recognition of various social structures, e.g. if the material only makes reference to Christian beliefs, learners whose families practice other religions will feel excluded.

110

16Score Total ×

1 2 3 4

1 2 3 4

1 2 3 4

1 2 3 4

4. Approach Weight = 5%

4.1 The learning and teaching strategies used in the LTSM are appropriate for the content i.e. there is a proper balance between pair, group, class and educator – led activities.4.2 There are various types of activities, e.g. discussions, debates, projects, etc to accommodate learners from different backgrounds, at different levels of development and with differing language proficiencies. 4.3 The various activities enhance learning and encourage learners to be active and constructive participants.

15

12Score Total ×

1 2 3 4

1 2 3 4

1 2 3 4

68

Page 69: SECTION ONE EXPLORATION OF THE STUDY AREA …

Rating scale:Very little = 1 Partially = 2 Mostly = 3 Very much so = 4

Circle the score that best describes the material/s.

7. Content Weight = 10%

7.1 The content is unbiased and free of prejudice. It should avoid gender stereotyping, racial inferences, offensive textual and visual material.7.2 The activities geared towards the attainment of the Assessment Standards correlate with the age and level of the learner. (The activities should be neither too simple nor too complex.) 7.3 There is a fair representation of different cultural groups in illustrations, names of characters and text.7.4 The content is accurate, up to date, coherent and conceptually sound.7.5 The content is based in the South African context and also including updated information where relevant.

110

20Score Total ×

1 2 3 4

1 2 3 4

1 2 3 4

1 2 3 41 2 3 4

8. Relationship between content, illustrations and activities Weight = 5%

8.1 The activities are sufficiently related to the text and/or illustrations that coincide with them.8.2 Content, illustrations and related activities address the same outcome. However, it is possible that the activities incorporate other Learning Outcomes through the principle of integration.8.3 The activities prompt the learners to participate in different ways to avoid monotony, e.g. a reading text should not always be followed by questions to be answered orally.

15

12Score Total ×

1 2 3 4

1 2 3 4

1 2 3 4

9. Language and Style Weight = 10%

9.1 The vocabulary used in the materials is clear, accessible and appropriate for target group. 9.2 New concepts and information are introduced in a clear and logical manner. 9.3 The language used does not discriminate against certain races. 9.4 The language used is sensitive to issues of gender. 9.5 The language used does not denigrate certain religions and cultures. 9.6 The language used does not encourage stereotyping. 9.7 The language caters for learners with a variety of language backgrounds and abilities

110

28Score Total ×

1 2 3 4

1 2 3 41 2 3 41 2 3 41 2 3 41 2 3 41 2 3 4

Rating scale:Very little = 1 Partially = 2 Mostly = 3 Very much so = 4

69

Page 70: SECTION ONE EXPLORATION OF THE STUDY AREA …

Circle the score that best describes the material/s.10 Layout and Design

Weight = 10%

The page layout is suitable in terms of:

10.1 clear headings and sub-headings. 10.2 clearly marked activities. 10.3 suitable print size and type. 10.4 adequate white space and margins. The manner in which the design features are used:

10.5 is appealing and attractive to the eye of the learner. 10.6 draws attention and holds interest for learners to interact continuously with the material.

110

24Score Total ×

1 2 3 41 2 3 41 2 3 41 2 3 4

1 2 3 41 2 3 4

11.Pictures and Illustrations Weight = 5%

11.1 The pictures and illustrations are relevant to the activities being dealt with, i.e. that pictures and illustrations are not there as gap-fillers or merely for decorative purposes. 11.2 The pictures and illustrations contain all the details mentioned in the activities 11.3 The pictures and illustrations are representative of different races and genders as well as people with special needs, where necessary and appropriate.

15

12Score Total ×

1 2 3 4

1 2 3 4

1 2 3 4

12. Educator Support Weight = 15%

12.1 There is clear reference and cross-reference to all the components of the package.12.2 The educator is encouraged to adapt material to suit his/her individual situation andcontext

The educators’ guide gives:

12.2 Brief overview of the curriculum.12.3 Guidance to the educator on how to facilitate activities.12.4 Guidance to the educator on the application of different assessment methods.12.5 Solutions to problems.12.6 Guidance to educators to appropriate methods and learning strategies

115

28Score Total ×

1 2 3 4

1 2 3 4

1 2 3 41 2 3 41 2 3 41 2 3 41 2 3 4

TOTAL SECTION A SCORE:

1. Add the scores to get a total

70

Page 71: SECTION ONE EXPLORATION OF THE STUDY AREA …

2. Write this score for section A on page 1

ADDENDUM B: TABLES AND GRAPHS

71

Page 72: SECTION ONE EXPLORATION OF THE STUDY AREA …

TABLE 1 Categories as determined by the focus group interviews (modified and

adapted from Maykut and Morehouse 1994)

CODES POSSIBLE CATEGORIESEvaluationhappens

with someassistance(independe

ntly)

Instrument is

the core

element for

evaluation

process

Instrument

encourages

collaborative

evaluation

Willingness among

educators to perform

the function as part of

their work

Inadequate support

to institutions

X

Instrument used by

different sections in

the GDE

X X

Instrument well

marketed within the

GDE

X X

Effective training

sessions

X

Instrument regularly

used

X

User friendly

instrument and well

structured

X

Instrument

addresses the needs

(including curriculum)

X

Instrument enables

the educator to

decide on LTSM

X

Instrument should be

improved

X

Instrument used in

collaboration

X

72

Page 73: SECTION ONE EXPLORATION OF THE STUDY AREA …

Instrument boosts

confidence in

evaluation

X

Evaluation is

indispensable

X

Table 2. Suggestions made by evaluators on changes that the GDE should

consider to effect on the instrument used for evaluation

Category Tallies Frequency

73

Page 74: SECTION ONE EXPLORATION OF THE STUDY AREA …

Update information on

front page of instrument

IIII III

08

Revised section on Notes

to the evaluators

IIII IIII IIII I I6

Maintain section for

evaluating readers,

games and charts

IIII IIII IIII IIII IIII IIII IIII

IIII I

41

Change section for

evaluating readers,

games and charts

IIII IIII IIII II 17

Lengthen instrument IIII II 07

Shorten instrument IIII IIII IIII I 16

No response/ no

changes suggested

IIII IIII IIII IIII IIII IIII IIII

IIII IIII IIII II

52

TOTAL 157

74

Page 75: SECTION ONE EXPLORATION OF THE STUDY AREA …

0

10

20

30

40

50

60

NUMBER OF RESPONDENTS

UPDATEFRONT

REVISE NOTES MAINTAINSECTIONS

CHANGESECTIONS

LENGTHENINSTRUMENT

SHORTENINSTRUMENT

NO RESPONSE

S1

RESPONSES

Graph 1: IMPROVEMENTS AS SUGGESTED BY EDUCATORS ON THE GDE EVALUATION INSTRUMENT

75

Page 76: SECTION ONE EXPLORATION OF THE STUDY AREA …

Table 3: Evaluators’ indication of how frequent they use the evaluation

instrument

Category Tallies FrequencyInstrument used once per

year

IIII IIII IIII IIII IIII III

28

Instrument used twice

per year

IIII I 06

Instrument used once per

quarter in a year

IIII IIII 10

Instrument used once per

month

IIII IIII 10

Instrument used more

often than the above

IIII IIII IIII 15

Educators do not use the

instrument or are not

sure how often they use

it

IIII IIII IIII IIII IIII IIII IIII

IIII IIII IIII IIII IIII IIII IIII

IIII IIII IIII III

88

TOTAL 157

76

Page 77: SECTION ONE EXPLORATION OF THE STUDY AREA …

28

610 10

15

88

0

10

20

30

40

50

60

70

80

90

NUMBER OFEDUCATORS

ANNUALLY SEMESTERLY QUARTERLY MONTHLY MORE NONE

FREQUENCY

Graph 2: EDUCATORS' UTILIZATION OF THE GDE LTSM EVALUATION INSTRUMENT

77

Page 78: SECTION ONE EXPLORATION OF THE STUDY AREA …

Table 4: How educators view complexity of using the LTSM evaluationinstrument

Category Tallies Frequency

Instrument easy to follow IIII IIII IIII IIII IIII III IIII

IIII IIII IIII III

53

Instrument difficult to use IIII IIII IIII IIII IIII IIII IIII I 36

Follow with assistance IIII IIII IIII 14Instrument is not usable

at all

IIII I O6

Educators not sure or do

not use the instrument at

all

IIII IIII IIII IIII IIII IIII IIII

IIII IIII II

47

No response I 01

TOTAL 157

78

Page 79: SECTION ONE EXPLORATION OF THE STUDY AREA …

0

10

20

30

40

50

60

NUMBER OF RESPONDENTS

INSTRUMENT COMPLEXITY

Graph 3: EDUCATORS EXPOSITION OF INSTRUMENT COMPLEXITY

EASY DIFFICULT ASSISTANCE NOT USABLE NO RESPONSE

79

Page 80: SECTION ONE EXPLORATION OF THE STUDY AREA …

ADDENDUM C: TEACHERS AS EVALUATORS OF LEARNING ANDTEACHING SUPPORT MATERIALS

As stated in Section 1, this work is focused on the evaluation of Learning and

Teaching Support Materials (LTSM) as opposed to textbooks. The reason is

centreed on the fact that textbooks are not the only but one type of LTSM.

However, other types of LTSM, like educators’ guides, learners’ guides and

workbooks, may be regarded as forms of textbooks. As specified by Davies

(2000:126), textbooks, in whatever form, are aimed at fulfilling three main

functions, namely ‘interpersonal’ (relationship between writer and reader),

‘ideational’ (selection of information) and textual (constructing a coherent

message). Evaluators, by means of evaluation instruments endeavour to

establish whether these functions are fulfilled by LTSM.

As required by the National Department of Education through the Revised

National Curriculum Statements (RNCS), publishers develop their LTSM in the

form of a submission which commonly consists of a Teacher’s Guide, Learners’

Work/ Activity book, and Reader. As implied earlier on, LTSM also include

puzzles, charts, audiovisual materials, games, magazines, newspapers and

another wide variety of materials. The aim of this study is to assess the

usefulness of the evaluation instrument that is used for selection of LTSM. An

inquiry will be made on how educators engage with the instrument.

Rationale for Evaluation of Learning and Teaching Support Materials

As a starting point, one needs to consider the meaning of the word ‘evaluate’.

The concept is synonymous with words like ‘assess’, ‘appraise’, ‘weigh’,

‘consider’,’ review’ ‘gauge’, ‘calculate’ ‘moderate’, ‘adjudicate’ ‘referee’ ‘umpire’

and ‘estimate’. Interestingly, the common goal or intention implied in all the

aforementioned terms is to find the value, worth, or usefulness of the subject.

This is what evaluators of materials are conscious about when evaluating LTSM.

To be more precise, evaluation of LTSM entails objective and intensive reading,

80

Page 81: SECTION ONE EXPLORATION OF THE STUDY AREA …

bearing in mind criteria as stipulated in an evaluation instrument; writing down

analytic points (strong and weak to determine whether the LTSM are

educationally sound); scoring and writing qualitative reports/ reviews about every

component of a publishers’ LTSM.

Commonly, upon completion of the evaluation process, education departments

select LTSM that meet specified minimum requirements and place those in a list

of approved LTSM. The list is eventually converted into a catalogue. GDE

names that ‘List of Approved Materials’. It is important to state that evaluation

occurs at different levels:outside the school at the Evaluation Centre and at the

school by individuals and teacher teams (e.g. LTSM committees within the

school). This implies that schools have a role to play in the selection of LTSM

considering the needs of their teachers, learners and the surrounding school

communities.

This serves as a complement for what schools and school communities agree on

as determined by and as a result of planning that is done at Macro (including

school and its community members), Meso (at the level of Heads of Departments

and their departmental educators) and Micro (involving the grade and learning

Area educator) level.

The objective for LTSM Evaluation and Review is to make a positive impact on

the effectiveness and efficiency of its provisioning, defined as having the right

materials at the least cost in the right institutions at the right time (Eastern Cape

Department of Education, 2004). In order to perform the task of evaluation of

LTSM successfully, the evaluator needs to be given preset criteria in order to

ascertain the parameters within which to function. The criteria are set up by the

relevant education department, in this regard Gauteng Department of Education

(GDE) through its LTSM unit. The criteria come in the form of a grade or Phase

– specific evaluation instrument that is adapted to the needs of the curriculum,

i.e. the Revised National Statement.

81

Page 82: SECTION ONE EXPLORATION OF THE STUDY AREA …

However, it is important to note that evaluation of LTSM is not the only activity in

an education departments’ process map but forms an important and

indispensable area. This is evident in Figure 1 and 2 as specified by models

used by the Eastern Cape Department of Education and GDE respectively.

Figure 2.1. Generic Building Blocks of a Supply Chain (The Eastern Cape

Education Department Model) Key Stages Identification of

requirementSelection andprocurement

Delivery & payment

Typical Building Blocks Assessing Stock in hand Identification ofsuitable suppliers andmaterials

Production

Determining requirementquantity and quality

Determining price Delivery

Assessing required levelof supply

Selection of price Distribution

Assessing resourceavailability

Ordering Payment

Inter-dependent OperationalDimensions

Typical People IssuesWho? Are they capable and willing? Are their mandates clear? Typical Systems and Process IssuesAre processes sufficient, cost effective and correctly sequenced? Arethey integrated? Do system arrangements support effective and efficientsupply and prevent leakages Typical Technology issues

Are appropriate technologies in place to support processes? Are they

used? Can they be improved?

Curriculum changes have necessitated provincial education departments to

replace LTSM that schools have been using and that has also prompted LTSM

evaluation and review processes that are provincially controlled. Gauteng

Department of Education is no exception to the rule. In order to accomplish the

task, provincial education departments have to design a criteria-bound and

curriculum compliant evaluation instruments. In view of the narrow gap between

82

Page 83: SECTION ONE EXPLORATION OF THE STUDY AREA …

the grades, the GDE designed an inclusive Foundation Phase and Intermediate

Phase evaluation instruments. However, due to complexity in the subsequent

Phases and grades and also the implementation plan of the national Department

of Education, a grade specific instrument may have to be designed and

implemented per grade. This has been the case with instruments that were

designed for curricula prior to the RNCS.

83

Page 84: SECTION ONE EXPLORATION OF THE STUDY AREA …

SendPODs forGDED/OEdusolutions forpayment

LTSM – Process Map

PROCESSReviewDevelop

material

(Proposal)

Evaluation

Expo

Requisitioning-GDE601 forstationery-GDE600 fortextbooks

Orderingbysection21.1.cschools(from theapprovedlist)

CapturingOrderingfor Non-section21.1.c

Orderingbysuppliers

Requestforprintingof LTSM

Printing(textbooks)Manufacturing(stationery)

Deliverytoschools

Distribution ofLTSM

Retrievalprocess

Schoolspayinvoiceas perPOD

RESPONSIBLEPERSON

DistributorsServiceProviders(SMMEs)

SLTSMCSGBEducators

LTSMEducators/SLTSMC

Printers/manufactures

Publishers –stationerytenderers

TextbooksuppliersStationerytenderers

GDED/O orServiceproviders

Section21.1.cschools

Non-Section21.1.cschools

H/OD/O

GDE(H/O)

GDE(H/O) orServiceProvider

PublishersAuthorMaterialdevelo

LTSM(manuscript)

List ofapproved/ RejectedMaterials

Individual reviewguide

Requisition (toD/O orServiceproviders)

Orders(tosuppliers)

Orders/sundry (tosuppliers)

Orders(topublishers)

PrintedLTSMandstationery

POD toschools

DistributedLTSM

Retrievedmaterial

Deliveryschedule

Procurement, Monitoring, Control and Reporting Stages (by EduSolutions) PRE-Procurement Stage (Evaluation and Selection)

Figure 2.2: The GDE LTSM Process Map

84

Page 85: SECTION ONE EXPLORATION OF THE STUDY AREA …

PRODUCT/OUTPUT

85

EXPO-samples-price list-Catalogue

PrintedandmanufacturedLTSMforschools

Page 86: SECTION ONE EXPLORATION OF THE STUDY AREA …

What is expected of an educator as an evaluator of LTSM? (An ideal LTSMEvaluator)

Based on the above, one may argue that the task of evaluating LTSM is

extremely significant as it is a means to ensure that the design features of the

National Curriculum Statement ( viz:Critical Outcomes, Assessment Standards

Learning Outcomes) are achieved. This would assist in ensuring that South

Africa ends up with competitive, world-class citizens.

The above-mentioned implies that education departments are compelled to take

evaluation as an indispensable, vital function. Logically it implies that the GDE

has to ensure that anyone who is given the task possesses the appropriate skills,

experience and attitude. This means that the process of selecting evaluators

has to be accurately executed. The GDE uses largely educators as evaluators

of LTSM and criteria that are used as a measure include, among other things,

prior experience, Learning Area experience and experience of teaching in the

grade.

As stated above, prior evaluation experience is considered when selecting

evaluators of materials to select materials for the entire Gauteng province. This

is based on the fact that the GDE is faced with the task of training all the

selected evaluators and, it goes without training that it is much easier to train a

group in which the majority of the participants have been trained before. Some

of the experienced evaluators are given positions of more responsibility at the

Evaluation Centre e.g. they are selected as chief evaluators. Besides for the

aforementioned, the likelihood is that the experienced evaluators will perform the

function better than a novice evaluator.

Likewise, the teacher’s Learning area/ programme experience is also regarded

as a vital criterion in the selection of an evaluator. All educators are expected to

have some experience in teaching in general. This refers to the so called

traditional or NATED-550 curriculum. However, other teachers have not had

86

Page 87: SECTION ONE EXPLORATION OF THE STUDY AREA …

training on contemporary curriculum issues ranging from the time of Curriculum

2005 to the current era of Revised National Curriculum Statements. Teachers

who have been trained and have taught in the learning area/ programme are a

preference. However, that does not mean that the rest will be marginalized but

they are given some special kind of attention in order to make up for their

inexperience should they be utilized. The same considerations are given with

the experience in teaching the grade.

87

Page 88: SECTION ONE EXPLORATION OF THE STUDY AREA …

ADDENDUM D: APPLICATION TO BE LTSM EVALUATOR

GAUTENG DEPARTMENT OF EDUCATIONCHIEF DIRECTORATE:CURRICULUM AND PROFESSIONAL DEVELOPMENT AND SUPPORT

APPLICATION TO BE APPOINTED AS AN EVALUATOR OF LEARNING AND TEACHINGSUPPORT MATERIALS (LTSM) AND LIBRARY RESOURCE MATERIALS (LRM)

PLEASE NOTE

1. Only the following persons may apply• Educators presently teaching in the Foundation Phase• University lecturers with OBE / NCS knowledge• Consultants of NGOs currently involved in OBE activities• District and Head Office Learning Area / Programme facilitators• Educators with experience in evaluation of Library and classroom materials

2. Applications must be submitted to the relevant District Office:EducationResources LSM facilitator by 25 June 2003

3. Please attach a certified copy of:• Your qualifications (Including Matriculation Certificate and Statements

of Results)• Your most recent salary advice reflecting the date (if employed by the

GDE)• A Review Report of a book you have recently read and evaluated

4. The selected applicants must be available for all meetings (even onWeekends)

Please complete only one character / number per block, or mark the appropriate block with an “X”

PERSONAL DETAILS

Identity number

PERSAL NO

Title Initials

88

Page 89: SECTION ONE EXPLORATION OF THE STUDY AREA …

First Names

Home Language

Surname

Postal Address

Postal code

Telephone Number (work)Telephone Number (Home)Fax NumberCell Number

Preferred Language for CommunicationEnglish

Afrikaans

ACADEMIC DETAILS (Completed Degrees / Diplomas only)

Academic QualificationsDegree / Diploma Institutions Major subject (s) Year completed

Professional / Technical QualificationsDegree / Diploma Institution Major subject(s) Year completed

Institution where employed

89

Page 90: SECTION ONE EXPLORATION OF THE STUDY AREA …

Centre Number

Present post held (e.g. Educator. HOD. Principal)

TEACHING EXPERIENCE

NAME OF SCHOOL INSTITUTION

POSITION HELD

SUBJECTS /LEARNING PROGRAMMES(AREAS) TAUGHT

GRADESe.g.1,2,3

PERIOD

FROM TO

EXPERIENCE IN EVALUATIONLEARNING AREA/ PROGRAMME / SUBJECT IN WHICH YEARS?

LEARNING PROGRAMMES OF CHOICE FOR EVALUATION (State three in order preference)

TEACHINGEXPERIENCE IN YEARS

1ST

LanguageofLearningandTeaching.

2nd

LanguageofLearningandTeaching

1.2.3.4.5.

I CERTIFY THAT THE ABOVE INFORMATION IS CORRECT AND ACCURATE AND ACKNOWLEDGE THAT ANYERRORS OF FALSE DECLARATION IN THIS APPLICATION WILL LEAD TO ITS DISQUALIFICATION.

_____________________________ ________________________ ____________________SIGNATURE OF APPLICANT PRINT NAME DATE

C. DECLARATION

90

Page 91: SECTION ONE EXPLORATION OF THE STUDY AREA …

1. THE PRINCIPAL /UNIT HEAD OF ________________________ SCHOOL/ OFFICCEDECLARE THAT ALL THE INFORMATION IN THIS APPLICATION IS CORRECT TOTHE BEST OF MY KNOWLEDGE AND THAT THE APPLICANT IS QUALIFIED IN ALLRESPECTS TO EVALUATE MATERIALS FOR THE LEARNING PROGRAMMEAPPLIED FOR.

NAME OF PRINCIPAL/ UNIT HEAD__________________

SIGNATURE OF PRINCIPAL / UNIT HEAD_________________

_________________________

CONTACT TEL NO.

__________________________DATE

SCHOOL DATE STAMP

________________________

BANK DEPOSIT DETAILS OF APPLICANT

NAME OF APPLICANT:_______________PERSAL NO:____________________NAME OF ACCOUNT HOLDER:_________________BANK:_________________________BRANCH:________________________ BRANCH CODE:_______________ACCOUNT NO:______________________TYPE OF ACCOUNT:____________________________

(Current, Transmission, Cheque, Savings etc)

For District Use Only

I hereby verify that:

1. The applicant is currently teaching in the Phase2. I recommend / do not recommend the applicant to be selected as evaluator.

NAME OF FACILITATOR/ DISTRICT LSM CO-COORDINATOR:_________________________________________-

SIGNATURE:_____________________DATE____________________

For Head Office Only

91

Page 92: SECTION ONE EXPLORATION OF THE STUDY AREA …

Application Accepted / Not Accepted

NAME OF PANEL MEMBER

SIGNATURE:_________________________

DATE:_________________________

92

Page 93: SECTION ONE EXPLORATION OF THE STUDY AREA …

ADDENDUM E: ABSTRACT FROM GDE LOETA POLICY DOCUMENT

GAUTENG DEPARTMENT OF EDUCATION

UMnyango WezeMfundo Lefapha la Thuto

Department of Education Departement van Onderwys

Draft LOETA PROSUREMENT SYSTEM POLICY

A. BACKGROUND

Prior to 1997, the GDE used a system to select Learning and Teaching Support

Materials (LTSM) on behalf of the educators in which educators were not involved. The

GDE realised that the system had some shortcomings and in order to rectify that

developed a system referred to as the ‘Open system’ in which educators had a role to

play in selecting LTSM. The ‘Open system’ implied that review, selection and evaluation

of LTSM took place at various levels (including school level). In addition, and also in line

with the principle of Self-Managing Institutions as rooted in the South African Schools’

Act, the ‘Open system’ allowed schools to commence to manage their own budgets,

make own choices on stationery (as opposed to pre-packs) and select textbooks that

are presented during the publishers’ marketing drive process in the form of exhibitions,

Schools-to-School visits and other means (i.e. from the open market).

Furthermore, the GDE introduced a Review Guide that is in the form of a collection of

review reports and which serves as an optional aid to the educator. The GDE therefore

stopped issuing a catalogue or any type of LTSM list for schools. It was hoped that

every educator would be trained on how to evaluate and select LTSM independently.

Such training would be backed by exhibitions and displays of publishers’ LTSM. The

main tool to assist the educator in evaluating the LTSM would be the Evaluation

Instrument.

93

Page 94: SECTION ONE EXPLORATION OF THE STUDY AREA …

In view of the above-stated shortcomings, the ‘Open system’ was put under debate to

subject it to observation, evaluation and review to test for its suitability, compatibility,

relevance and usefulness. Consultation was done with various Head Office and District

units (in particular, LTSM and Curriculum units). These debates were also backed by

comparative work that was done with some provinces. The outcome was the

replacement the system with a ‘Limited Open (equivalent to approval) System’ (LOETA

system). In the system, the educators still have an active role to play but have to select

their classroom materials from a list of approved LTSM. This also implies that the GDE

will now take a decision on the criteria for selection based on the curriculum

requirements, decide on a number of submissions per Learning Area/ Programme,

produce a list of approved LTSM and advise the educators accordingly. Only materials

that were approved will be exhibited, procured and distributed to the schools.

94

Page 95: SECTION ONE EXPLORATION OF THE STUDY AREA …

B. LTSM PROCESSES AND CONCEPTS WITHIN THE LOETA SYSTEM

1. DEVELOPMENT OF LTSM

This is the initial step in the LTSM process map. Different persons including the

Learning Area educator, District and Head Office officials and the publishers’ authors

do LTSM development. The GDE also has an important role to play in the

development of educators on how to develop materials. This assists in preparing

the educators for the evaluation, selection and review and utilization of LTSM

processes that will be explained below. This step is characterised by, among other

things, the actual writing of printed materials by authors, development of educational

equipment (e.g. puppets, toys, puzzles etc). However, it is important to note that

the emphasis of the procurement process is to be discussed hereunder.

2. SUBMISSION OF LTSM BY PUBLISHERS

Once the development of LTSM discussed above concludes, the publishers desire to

supply schools with their materials (i.e. they require the GDE and its schools to

procure the LTSM). As a pre-requisite the GDE needs to evaluate and possibly

review the LTSM prior to the procurement process. Every year, the GDE sets a date

on which all the publishers are required to submit LTSM that they desire to supply to

the schools.

Coupled with the submission date is the Submission Specification memorandum.

The Submission statistics include Typing specifications (font size, font style, program

(e.g. MS Excel) and specification on software to be used), Packaging specifications

(sorting by language and learning area, the contents and number of sets per

submission- normally three and cover details to be stated on the cover page per set.

Furthermore, the specifications memorandum specifies the submission fees, which

are normally charged per submission – The GDE uses its discretion whether to

charge per submission or per Learning Area/ Programme depending on a variety of

variables including the total number of submissions, number of implementing

grades, etc. The submission amount will always be reviewed and decided by the

GDE on an annual basis. The funds are credited to the Gauteng Education

Development Trust account. These funds are used to cater for the professional and

administrative costs of the evaluation process.

3. EVALUATION, REVIEW AND SELECTION OF LTSM

95

Page 96: SECTION ONE EXPLORATION OF THE STUDY AREA …

The three processes of Evaluation, Review and selection of LTSM as defined by the

GDE are interdependent processes and serve as a pre-requisite for LTSM

procurement. Depending on a number of possible causative factors, the GDE

possesses the right to use its prerogative in deciding whether or not to undertake

any one of the three steps. However, the definition below is meant for the purpose

of indicating the common understanding of the concepts among all users and

participants within the GDE

EVALUATION OF LTSM

During the Evaluation stage, LTSM that have been received from the publishers are

evaluated for appropriateness for the grade or and for usability in the classroom.

Panels of evaluators who have deep knowledge of the curriculum, who have/ are

ready to implement it and are competent in using evaluation instruments and

statistically designed score sheets are carefully selected and appointed on a

temporary basis. These evaluators are selected from the larger South African

community of professionals and may include school educators, education

practitioners, higher education educators/ lecturers and others.

The publishers’ materials are scored using a quantitative approach by means of a

score sheet. Materials that achieve more than the minimum score are put on an

‘Approved Materials List’. In the ‘Approved Materials List’, the choice will still be

wide, as a number of titles per Learning Area will be listed. Materials with minimal

errors are put on a ‘Conditionally Approved Materials List’ and the publishers are

given comments on the materials’ weaknesses and a re-submission date. Materials

that are far less than the minimum are put on a ‘Rejected List’ and are totally not

considered. However, the GDE may announce an appeal date during which

publishers whose materials have been rejected may appeal. The date should be

seven days after the publishers have been advised of the rejection of their materials.

A reasonable amount of time will be given for the publishers to rectify/ rewrite their

materials and resubmit. Announcement will then be made to all the publishers and

the final ‘Approved Materials List’ is produced and locked. The locking of the system

implies that no further entries will be added or removed from the list. Hard copies

and write-protected CDs will be sent to all the GDE institutions and offices prior to

the Exhibition period.

96

Page 97: SECTION ONE EXPLORATION OF THE STUDY AREA …

3.2 REVIEW OF LTSM

The Review process is characterised by the use of mostly qualitative Evaluation

instruments to write up reports on reviews done on materials. Only materials that were

evaluated in the last stage and put on the ‘Approved Materials List’ will be reviewed.The review reports are done at different levels:first by individual reviewers, secondly the

review reports are put together into a Composite report. The composite report is

checked and verified by the team leader, coordinator and submitted to the GDE Quality

assurance team. Once found to be of a reasonable quality the reports are listed and put

together into a compilation known as the Review Guide. The Review Guide is released

in both the form of a hard copy and CD.

The Review of LTSM is not only centrally done but is also done at schools level for

various reasons at different intervals. The schools educators and School Learning and

Teaching Support Materials Committees (SLTSMCs) are invited to exhibitions/ EXPO

that are organized by Head Office and District Offices. During the exhibitions the

educators are given the opportunity to view publishers’ materials; use their Evaluation

Instrument and Review Guide to read more on the materials that have been reviewed;

and select materials that they desire to procure for their schools. However, that is not

the final step, as the SLTSMCs will also nominate a selection committee that will use the

Evaluation Instrument to select from the ‘Approved Materials List’. Furthermore, the

school may suspect some materials to be outdated or obsolete. Before deciding on

transfer of the materials they need to use the Evaluation instrument and compile a

report stating why the materials are declared obsolete.

3.3 SELECTION OF MATERIALS

All the participants do the selection of LTSM at the schools level. These participants

include grade educators, Learning area/ programme educators, Heads of Departments,

the School Management teams, the SLTSMCs and the School Governing Bodies (SGB).

Once a decision has been finalized, the SLSMC puts together requisitions and orders

that will be explained in the next sections of this document.

97

Page 98: SECTION ONE EXPLORATION OF THE STUDY AREA …

4. EXHIBITIONS

In order to give the publishers’ materials maximum exposure and to ensure that all the

educators are aware of and explore the LTSM that are available in the market, the GDE

organizes exhibitions or LTSM EXPO. The EXPO is held at a central venue for a

blocked period of about seven days. The advantage of a central EXPO is that

publishers do not have to move on a daily basis from district to district. That saves on

valuable time and minimizes expenses. Furthermore it will enable publishers to bring a

wider variety of available materials across the grades. Monitoring the EXPO becomes

more efficient and sufficient.

During the EXPO, over and above the display of materials, seminars are held. Topics of

educational value are to be discussed. Presentations are to be done by various

stakeholders including publishers, suppliers, distributors, service providers, higher

education institution academics, independent intellectuals, publishers’ and bookseller’s

associations, education officials and officials from other state departments. Stalls will be

hired out to the exhibitors at a fee negotiated with all the stakeholders and an events

management service provider will be appointed to market, organize and finalize the

event.

98

Page 99: SECTION ONE EXPLORATION OF THE STUDY AREA …

5. REQUISITIONING AND CAPTURING The requisitioning process takes place both in Section 21.1.c and Non- Section

21.1.c schools. However, the process happens at two different levels at both the

school types. In Section 21.1.c schools it is an internal schools process in which

departments or Learning Areas requisition what they require to the SLTSMC who in

turn forward the requisitions to the School Management Team (SMT) and the SGB

for final approval.

In Non- Section 21.1.c schools the process referred to above initiates inside the

school but is concluded at the District Office. Once Requisitions are recommended

and signed, Non-Section 21.1.c schools forward the relevant forms to District office

or service provider (if appointed as this is regarded as a Non-Core function and the

GDE may decide to outsource the function). The District office/ Service provider

therefore scrutinizes, checks, verifies the forms for correctness and to ascertain

whether the schools have not exceeded the budget. After that the District office/

Service provider captures the requisitions on a specific/ compatible accounting

system. Once capturing is done, the next step, ordering takes place.

6. ORDERING

As mentioned in 4 above, Section21.1.c schools prepare and place their orders

directly with textbook distributors/ suppliers and stationery suppliers. The

Section21.1.c schools together with the distributors and suppliers need to sign terms

of agreement to ensure that deliveries and payments are done as per the

expectations of both the parties. Suppliers therefore place their orders with

publishers

As stated above, district offices/ service providers place orders on behalf of Non-

section21.1.c schools with textbook suppliers and stationery tenderers. In the

absence of the latter orders are placed with the service providers who in turn

place the orders with publishers. Upon receiving the orders publishers up

together composite orders to request the printers to perform the printing works.

The average duration the print run is eight weeks after which the printers deliver

to the publishers. In the case of stationery orders are placed with stationery

tenderers who in turn place the orders with stationery manufacturers. The

99

Page 100: SECTION ONE EXPLORATION OF THE STUDY AREA …

stationery manufactures are therefore expected to deliver directly to the

tenderers.

7. PRINTING

The printing of media/ printed LTSM (books, textbooks) is a lengthy period and

the GDE considers that in determining delivery dates. Publishers place an order

with printers. Averagely an allowance of eight weeks is given for the process.

On conclusion of printing, the printers and manufacturers deliver LTSM.

8. DELIVERIES AND PAYMENTS

The deliveries process presents itself as a reversal of the ordering process.

Printers/ stationery manufacturers deliver LTSM to the publishers who in turn

deliver to the suppliers/ distributors/ service providers who also deliver to

schools. When deliveries are done to schools, a delivery advice/ Proof of

Delivery (POD) is signed and stamped by the principal of the school. A copy is

left for the school’s records.

Schools are always urged to check the contents of the deliveries and reconcile

those with their original requisition forms. Non- Section 21.1.c schools are

required to present copies of their PODs to the District office or appointed

service provider for payment purposes. Section 21.1. c schools, upon checking

and finding the LTSM satisfactory, process and make the payment to the

suppliers/ distributors. The service provider/ district office will also pay the

supplier/ distributors per POD.

9. STAKEHOLDERS’ PARTICIPITATION

100

Page 101: SECTION ONE EXPLORATION OF THE STUDY AREA …

The GDE has established a forum in which all the stakeholders are invited to

make educationally sound contributions to the GDE education system. The

stakeholders include, among others, suppliers of stationery, distributors of

textbooks, publishers’ associations, booksellers associations, independent

publishers, academics, educators, etc. A stakeholders’ code of conduct has

been developed in consultation with stakeholders and is subject to public

scrutiny.

The purpose of the code of conduct includes managing and controlling activities

of independent suppliers and distributors some of whom access the GDE

Section 21.1.c schools directly. All the stakeholders are required to sign and

acknowledge the existence of the code of conduct. Stakeholders’ Forum

meetings are arranged for and coordinated by the GDE in consultation with

stakeholders.

101

Page 102: SECTION ONE EXPLORATION OF THE STUDY AREA …

ADDENDUM F: SURVEY QUESTIONNAIRE

SURVEY QUESTIONAIRE

NB:When completing this questionnaire, you are requested to:

• Please answer the questions with honesty• Answer all the applicable questions• Write legibly• Add whatever information that you deem important but was not asked for on

the margins• Avoid using your name as your identity is not required

Date:____________________

SECTION A:PERSONAL INFORMATION

Mark the appropriate block with an X or fill in the required information

Age (in years):

Gender: M F

Race: BLACK WHITE COLOURED INDIAN OTHER (specify)

Teaching experience (in years)

Location of school: Township Sub-urban Inner city Farm

SECTION B:EXPERIENCE IN EVALUATION

102

Page 103: SECTION ONE EXPLORATION OF THE STUDY AREA …

1. I have...Never seen the instrument beforeSeen but not used the instrumentOnly seen my colleagues use the instrument before Seen and used the instrument

2. I learnt about the instrument in the following way …

Through a workshop by the education departmentBy seeing my colleagues using it At the provincial Evaluation CentreFrom a close friend of mine who is a teacher

Note:Only respond to question 3 below if you have received training/ workshopby Gauteng Department of Education.

3. After receiving training on the utilization of the Evaluation Instrument I felt asfollows:

knowledgeable and confident to independentlyevaluate materialsCould try to evaluate but with a few doubts I have gained some knowledge but need more trainingAttending the training was not worthwhile

4. I have used the instrument intended for evaluating materials in the ….…..Phase

Foundation PhaseIntermediate PhaseSenior PhaseFurther Education and Training

5. I use the instrument …

Once per year Once in six monthsOnce per quarter in a yearOnce per month

103

Page 104: SECTION ONE EXPLORATION OF THE STUDY AREA …

More often than above

SECTION C:ACTUAL ENGAGEMENT WITH THE INSTRUMENT

1. The information that an evaluator is required to complete in the instrumentis….

SufficientAlmost sufficientLacks important detailsTotally insufficient

2. The way the procedure for evaluating materials is explained …

I can easily followI follow with some struggle I cannot follow without assistance from othersI cannot follow altogether

3. The instrument addresses the assessment qualities of the book…

SufficientlyVery wellModeratelyPoorlyInsufficiently

4. The way in which the instrument addresses the Teaching Approach used

by authors in their books (e.g. group, pair, class, and project)…

Focuses mostly on oneFocuses on twoFocuses on threeFocuses on all

5. The naming of the different sub-sections…

Does not capture all the relevant conceptsCaptures the relevant concepts to some extentAlmost captures all the relevant conceptsSufficiently captures all the relevant concepts

104

Page 105: SECTION ONE EXPLORATION OF THE STUDY AREA …

6. When using the instrument I…

Always find a repetition of some statementsOften find a repetition of some statementsSometimes find a repetition of some statementsNever find a repetition of any statements

7. The way the instrument is structured, I …

Am able to identify and eliminate inappropriate materialsI to a limited extent can differentiate between appropriateand inappropriate materialsI find it difficult to distinguish between appropriate andinappropriate materials I am unable to distinguish between appropriate andinappropriate materials

8. The instrument makes provision for

For some types of materialsFor all types of materialsOnly for prominent types of materials typesOnly for materials not intended for the said grade

9. If I were to improve on the instrument, I would do the following:

Update information on the front pageRevise the section called ‘notes to the Evaluator’ Effect changes but maintain sections for evaluatingreaders, games and charts Completely change the sections for evaluating readers,games and charts Lengthen the entire instrumentShorten the entire instrument

10.Overall, I think the instrument:

Is completely objective and fair to the publishersIs almost objective

105

Page 106: SECTION ONE EXPLORATION OF THE STUDY AREA …

Is almost subjectiveIs completely subjective and unfair to the publishers

ADDENDUM G: Interview transcripts

FOCUS GROUP INTERVIEWS

PLACE: ALBERTON TEACHERS CENTRE

DATE: 15 JULY 2004

CLUSTER NUMBER 1

RESEACHER: Colleagues, I would like us to take a discussion on the

process of evaluation and selection of LTSM. Precisely, the reason for me

to decide to select this particular group was based on your prior knowledge,

and experience.

JEREMY: Tell us exactly what areas you would like us to discuss.

RESEACHER: Oh well! This tells me that you have a lo to share with us.

What are the experiences that you have had as a participant in the process of

LTSM evaluation and selection?

RACHAEL: I know he has a lot to tell us because I was introduced by him

to the process. In fact, he increased my interest in the process of LTSM

evaluation.

106

Page 107: SECTION ONE EXPLORATION OF THE STUDY AREA …

JEREMY: It is actually a very complicated process when you are not given

support.

MAXWELL: Especially the so called District officials. At times I think

they are abdicating their responsibilities because they leave us without

support at all.

RESEACHER: What kind of support maybe?

SIZWE: For instance training us, giving us instruments and also feedback.

RESEACHER: Okay, I hear you... At what levels do the District officials

involve you as educators in the process?

HLOMU: They advise us to do that even at school level. In this case you

need to consult with other teachers via the LTSM committee to decide on

the kind of material to be selected for you classroom.

RACHAEL: Don’t forget the governing bodies. They will kill you!

JOAN: It is true they like to be involved but they need more support than us.

RESEACHER: Besides support from officials who else can give you the

necessary support.

107

Page 108: SECTION ONE EXPLORATION OF THE STUDY AREA …

JEREMY: I think we do a wonderful job when we work in our teams at

school and at provincial Evaluation Centres. Your colleagues give you

support, assurance and confidence.

RESEACHER: You have spoken about all the stake holders but have not

mentioned school management team members.

ELIZABETH: Forget about those. What they know is to hide away

circulars and memoranda that are sent by the department.

RESEACHER: Jeremy spoke about an Evaluation Centre. Do you think

that schools can do without having to go to the Evaluation Centre?

HLENGI: Yes we can because we do most of the work ourselves.

HLOMU: No! Those people give us a lot of guide lines. Without their

guidance our processes could not take off the ground.

SIZWE: In particular they supply us with the evaluation instrument.

Besides, teacher can be rowdy in the absence of authorities.

RESEACHER: How often do you think the evaluation process should take

place?

HLENGI: To be honest only when RNCS is introduced in a phase.

108

Page 109: SECTION ONE EXPLORATION OF THE STUDY AREA …

ELIZABETH: I definitely do not agree. I have been a team leader at the

centre and when we were trained it became evident that this process is

indispensable and needs to take place once per year.

RESEACHER: Do you get paid and how do you feel about that?

JEREMY: (pointing at Elizabeth) this one is supposed to support the

process because she scoops a lot of money during evaluations.

ELIZABETH: (In response) not necessarily for that purpose! Even if I am

not paid I would still do it. I actually think that it should be compulsory to

all the teachers.

HLOMU: You are right! Many of us think that the process would be more

effective if it takes place for several occasions in a year.

RACHAEL: Maybe they should establish a post for LTSM head of

department.

RESEACHER: Do you all think that the process should form part of your

work?

HLENGI: Over my dead body!

JEREMY: I think Rachael was right. This is a significant process for all the

stake holders in education.

109

Page 110: SECTION ONE EXPLORATION OF THE STUDY AREA …

RESEACHER: Thank you very much for availing yourselves to this

discussion. May I request you to avail yourselves to anyone who might

need an opportunity to utilize your knowledge and experience?

ALL: Thank you very much!

FOCUS GROUP INTERVIEWS

PLACE: ALBERTON TEACHERS CENTRE

DATE: 16 JULY 2004

CLUSTER NUMBER 4

RESEACHER: Colleagues, I would like us to take a discussion on the

process of evaluation and selection of LTSM. Precisely, the reason for me

to decide to select this particular group was based on your prior knowledge,

and experience. Do you think there is an improvement on the quality of

materials that are used in our classrooms and what is the cause?

BAFANA: Based on my experience as an educator, member of the materials

committee and evaluator of materials, I would think that the GDE is moving

towards the right direction. In the past we were excluded from the process

of taking decisions as to what kind of materials we wished to utilize to

develop our children. Materials that we didn’t desire were ordered on our

behalf. We were not given the opportunity to accept or reject. That was the

110

Page 111: SECTION ONE EXPLORATION OF THE STUDY AREA …

shame of Bantu education. We were forced to sign for and receive

undesired materials.

MARY: Although there are still some loopholes in the new system of

evaluating materials, we now feel as professionals who are respected as

opposed to being puppets.

RESEACHER: What kind of loopholes?

MARY: I am just hoping that there will be a time when the instrument that

we use for evaluation will assist the educator to select the materials that

don’t have any defects or weaknesses that could hamper or negatively affect

the teaching and learning processes in our schools. For now I think the

materials that we chose for ourselves still have some limitations.

ZANELE: I don’t squarely put the blame on the system that we use to

evaluate materials. The problem could be the non – availability of

appropriate materials from the publishing houses. Let us not blame the

department for poor materials in our systems but writers, developers and

publishers of materials.

ZANDILE: My worry is that the type pf materials that are available are

mainly text books. What about puzzles, charts, CDs, atlases, Dictionaries

and audiovisuals, etc.

RESEACHER: Do you think there is a special type of instrument to measure

the usefulness of the variety of materials?

111

Page 112: SECTION ONE EXPLORATION OF THE STUDY AREA …

BIGBOY: Presently there isn’t.

SANDILE: No Bigboy I think you are making a mistake when I was at the

Evaluation Centre, some publishers had submitted a wide variety of

materials including software. The challenge was to evaluate.

SIMON: What did you do?

ZANELE: I still recall our team was mainly given such materials and we

were directed to use a section of the instrument that was designed

particularly to evaluate such materials.

RESEACHER: Besides the actual evaluation, what did you enjoy or find to

be boring about being part of the evaluation group and to be at the

Evaluation Centre?

ZANDILE: The spirit of togetherness! To be taken care of by other

educators, listen to their advises , be criticesed, make a contribution,

develop ideas, write reports and await payment (laughing )

RESEACHER: Now that you have spoken about the payment of evaluators,

would you be willing to, at some point, participate in textbook evaluation

without remuneration?

SIMON: No ways, remember that as educators we receive meager funds in

the form of payments. I am therefore unwilling to work for no pay.

112

Page 113: SECTION ONE EXPLORATION OF THE STUDY AREA …

SANDILE: Remember that it is your calling. Besides, it is within the

principles of democracy that we should have a contribution to make in our

profession. We need to be proud by knowing that we were part of the

selection process. However, we will feel more secured and protected if we

have the backing and support of our line functionaries – in particular if our

principals could refrain from blockading the insemination of information .

RESEACHER: How could this is achieved?

SIMON: The best way is for the department of education to strengthen its

structures and also by means of consultative forums.

RESEACHER: I would like to thank you all for availing yourselves to this

short discussion.

ALL: You are welcome.

113

Page 114: SECTION ONE EXPLORATION OF THE STUDY AREA …

FOCUS GROUP INTERVIEWS

PLACE: ALBERTON TEACHERS CENTRE

DATE: 16 JULY 2004

CLUSTER NUMBER 3

RESEACHER: Colleagues, I would like us to take a discussion on the

process of evaluation and selection of LTSM. Precisely, the reason for me

to decide to select this particular group was based on your prior knowledge,

and experience. Who would like to start and share with us the

understanding of that process?

JOHNSON: That is a process that takes place when we are asked to go to

the exhibitions to represent our schools. We often go there on an afternoon

and are given samples of material to view and to go and share with our

colleagues at school. This is followed by the requisitioning of materials.

JOY: I don’t think the question relates to that.

114

Page 115: SECTION ONE EXPLORATION OF THE STUDY AREA …

RESEACHER: In fact the processes explained by Johnson do form part of

the evaluation and selection of LTSM. But the process starts at an earlier

stage.

PATRICK: It is when the department receives loads and loads of materials

and educators are requested to partake in the process of evaluating the

LTSM.

VICTOR: I have been to the Evaluation Centre and I do not wish to go back

there.

RESEACHER: Why?

VICTOR: Evaluators of materials work under severe pressure. There is no

time breathe. They divide you into teams and give you an unreasonable

amount of work.

WINNIE: Each team is given a minimum of ten submissions and is expected

to sit down and evaluate their ration without a break. Mind you this has to

happen over weekends.

MANDLA: Only three to four weekends.

RESEACHER: Do they assist you to do that.

JOY: Only by means of an evaluation instrument.

115

Page 116: SECTION ONE EXPLORATION OF THE STUDY AREA …

RESEACHER: Tell me about the instrument.

VICTOR: The instrument is grade – specific. It consists of questions that

are grouped according to themes and sub-themes. The questions are scored

according to either a three or five point scale. These questions assist the

evaluator on deciding whether or not a book is subjective, racist, gender-

insensitive, qualifies, and complies with the curriculum and so on.

SALLY: In addition the instrument has questions that deal with the physical

aspect. It is to test whether the book will last for a long period without

getting damaged. Also to determine whether the book is worth the price

that the publishers are asking for.

RESEACHER: Are you able to use that instrument without assistance?

VICTOR: If you have had some prior training you only need little

orientation to be able to cope with the task of evaluating materials for the

particular grade in the particular year.

KHAYA: Some of us have had an opportunity of being trained on several

occasions for different phases and grades. However, I have learned that

each year the instrument is up graded and is nearing perfection. That makes

it even simpler to use at the Evaluation Centre and at our schools.

SALLY: In fact I feel much freer to use the tool in the absence of

authorities. Under supervision, I feel threatened, tense and even lose

composure.

116

Page 117: SECTION ONE EXPLORATION OF THE STUDY AREA …

VICTOR: I also think so because when you are alone or at least with your

peers you are able to give yourselves autonomy, modify the instrument and

convert the quantities to qualities. In this situation you are able to discuss

with your colleagues some issues that would not have catered for by the

instrument.

RESEACHER: Do you think the department enables you as teachers to own

the process of evaluation and selection

JOHNSON: In the past the teacher was given very little opportunity to

partake in the process. We felt left out however, now we are given a certain

amount of control. When I was at the Evaluation Centre I wished that could

every month or at least once per quarter.

RESEACHER: I think we have covered a remarkable amount of the

intended outcomes of this participative discussion. May I thank you for

your time and active involvement in the process

117

Page 118: SECTION ONE EXPLORATION OF THE STUDY AREA …

FOCUS GROUP INTERVIEWS

PLACE: ALBERTON TEACHERS CENTRE

DATE: 16 JULY 2004

CLUSTER NUMBER 4

RESEACHER: Colleagues, I would like us to take a discussion on the

process of evaluation and selection of LTSM. Precisely, the reason for me

to decide to select this particular group was based on your prior knowledge,

and experience. Do you think there is an improvement on the quality of

materials that are used in our classrooms and what is the cause?

BAFANA: Based on my experience as an educator, member of the materials

committee and evaluator of materials, I would think that the GDE is moving

towards the right direction. In the past we were excluded from the process

of taking decisions as to what kind of materials we wished to utilize to

develop our children. Materials that we didn’t desire were ordered on our

behalf. We were not given the opportunity to accept or reject. That was the

118

Page 119: SECTION ONE EXPLORATION OF THE STUDY AREA …

shame of Bantu education. We were forced to sign for and receive

undesired materials.

MARY: Although there are still some loopholes in the new system of

evaluating materials, we now feel as professionals who are respected as

opposed to being puppets.

RESEACHER: What kind of loopholes?

MARY: I am just hoping that there will be a time when the instrument that

we use for evaluation will assist the educator to select the materials that

don’t have any defects or weaknesses that could hamper or negatively affect

the teaching and learning processes in our schools. For now I think the

materials that we chose for ourselves still have some limitations.

ZANELE: I don’t squarely put the blame on the system that we use to

evaluate materials. The problem could be the non – availability of

appropriate materials from the publishing houses. Let us not blame the

department for poor materials in our systems but writers, developers and

publishers of materials.

ZANDILE: My worry is that the type pf materials that are available are

mainly text books. What about puzzles, charts, CDs, atlases, Dictionaries

and audiovisuals, etc.

RESEACHER: Do you think there is a special type of instrument to measure

the usefulness of the variety of materials?

119

Page 120: SECTION ONE EXPLORATION OF THE STUDY AREA …

BIGBOY: Presently there isn’t.

SANDILE: No Bigboy I think you are making a mistake when I was at the

Evaluation Centre, some publishers had submitted a wide variety of

materials including software. The challenge was to evaluate.

SIMON: What did you do?

ZANELE: I still recall our team was mainly given such materials and we

were directed to use a section of the instrument that was designed

particularly to evaluate such materials.

RESEACHER: Besides the actual evaluation, what did you enjoy or find to

be boring about being part of the evaluation group and to be at the

Evaluation Centre?

ZANDILE: The spirit of togetherness! To be taken care of by other

educators, listen to their advises , be criticesed, make a contribution,

develop ideas, write reports and await payment (laughing )

RESEACHER: Now that you have spoken about the payment of evaluators,

would you be willing to, at some point, participate in textbook evaluation

without remuneration?

SIMON: No ways, remember that as educators we receive meager funds in

the form of payments. I am therefore unwilling to work for no pay.

120

Page 121: SECTION ONE EXPLORATION OF THE STUDY AREA …

SANDILE: Remember that it is your calling. Besides, it is within the

principles of democracy that we should have a contribution to make in our

profession. We need to be proud by knowing that we were part of the

selection process. However, we will feel more secured and protected if we

have the backing and support of our line functionaries – in particular if our

principals could refrain from blockading the insemination of information .

RESEACHER: How could this is achieved?

SIMON: The best way is for the department of education to strengthen its

structures and also by means of consultative forums.

RESEACHER: I would like to thank you all for availing yourselves to this

short discussion.

ALL: You are welcome.

121