Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms...

50
Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore David Hogan*, Melvin Chan, Ridzuan Rahim, Dennis Kwek, Khin Maung Aye, Siok Chen Loo, Yee Zher Sheng and Wenshu Luo National Institute of Education, Singapore By any measure, Singapore’s educational system has generated an extraordinary record of achieve- ment over the past two or three decades. In this article, we report on one key component of a broader three year investigation into why Singapore has done so well, and explore the logic, strength, resilience and limits of the underlying pedagogical model and policy framework that have helped secure this record of achievement. Specifically, we draw on data we collected in 2010 to ana- lyze the pedagogical organization of four theoretically specified ‘models’ of instructional strategy– traditional instruction, direct instruction, teaching for understanding, and co-regulated learning strategies–in Secondary 3 mathematics and English. In the course of our analysis, we develop three arguments. The first is the single-minded performative orientation of instructional practices gener- ally–and instructional strategies specifically–in Singaporean classrooms that rarely deviated from a logic of curriculum coverage, knowledge transmission and assessment. Second, while we found sub- stantial evidence of a pervasive performative orientation to instruction, we also found that teachers in Singapore draw from a variety of instructional perspectives in ways that reflect a pragmatic, instrumental fit-for-purpose approach and broader performative orientation. Third, we found that the national high stakes assessment system, by virtue of its considerable institutional authority, both shaped the pattern of instructional practice at the classroom level and constrained opportunities for instructional improvement. In the conclusion, we review related findings from the research program on the impact of instructional practice on student achievement in Singapore. Introduction By any measure, Singapore’s educational system has generated an extraordinary record of achievement over the past two or three decades. Not the least of these achievements has been its record in international assessments, including TIMMS and PISA (Table 1). It is now recognized as one of the leading educational systems in the world and the object of envy and emulation. In this article we report on one key component of a broader three year investigation into why Singapore has done so well and explore the logic, strength, resilience and limits of the underlying pedagogical *Corresponding author: National Institute of Education, 1 Nanyang Walk, Singapore 637616, Singapore. Email: [email protected] Review of Education Vol. 1, No. 1, February 2013, pp. 57–106 DOI: 10.1002/rev3.3002 © 2013 British Educational Research Association

Transcript of Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms...

Page 1: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

Assessment and the logic of instructional

practice in Secondary 3 English and

mathematics classrooms in Singapore

David Hogan*, Melvin Chan, Ridzuan Rahim, Dennis Kwek,Khin Maung Aye, Siok Chen Loo, Yee Zher Sheng andWenshu LuoNational Institute of Education, Singapore

By any measure, Singapore’s educational system has generated an extraordinary record of achieve-

ment over the past two or three decades. In this article, we report on one key component of a

broader three year investigation into why Singapore has done so well, and explore the logic,

strength, resilience and limits of the underlying pedagogical model and policy framework that have

helped secure this record of achievement. Specifically, we draw on data we collected in 2010 to ana-

lyze the pedagogical organization of four theoretically specified ‘models’ of instructional strategy–

traditional instruction, direct instruction, teaching for understanding, and co-regulated learning

strategies–in Secondary 3 mathematics and English. In the course of our analysis, we develop three

arguments. The first is the single-minded performative orientation of instructional practices gener-

ally–and instructional strategies specifically–in Singaporean classrooms that rarely deviated from a

logic of curriculum coverage, knowledge transmission and assessment. Second, while we found sub-

stantial evidence of a pervasive performative orientation to instruction, we also found that teachers

in Singapore draw from a variety of instructional perspectives in ways that reflect a pragmatic,

instrumental fit-for-purpose approach and broader performative orientation. Third, we found that

the national high stakes assessment system, by virtue of its considerable institutional authority, both

shaped the pattern of instructional practice at the classroom level and constrained opportunities for

instructional improvement. In the conclusion, we review related findings from the research program

on the impact of instructional practice on student achievement in Singapore.

Introduction

By any measure, Singapore’s educational system has generated an extraordinary

record of achievement over the past two or three decades. Not the least of these

achievements has been its record in international assessments, including TIMMS

and PISA (Table 1). It is now recognized as one of the leading educational systems in

the world and the object of envy and emulation. In this article we report on one key

component of a broader three year investigation into why Singapore has done so well

and explore the logic, strength, resilience and limits of the underlying pedagogical

*Corresponding author: National Institute of Education, 1 Nanyang Walk, Singapore 637616,

Singapore. Email: [email protected]

Review of EducationVol. 1, No. 1, February 2013, pp. 57–106

DOI: 10.1002/rev3.3002

© 2013 British Educational Research Association

Page 2: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

model and policy framework that have helped secure this widely admired and envi-

able record of achievement.

Specifically, we draw on data we collected in 2010 to analyze the pedagogical orga-

nization of four theoretically specified ‘models’ of instructional strategy—traditional

instruction, direct instruction, teaching for understanding, and co-regulated learning

strategies—in Secondary 3 mathematics and English. We do not report, however, on

other key features of instructional practice in Singapore—the design and enactment

of instructional tasks, classroom organization, interaction and talk, the use of high

leverage instructional strategies (including checking for prior knowledge, the commu-

nication of learning goals and performance standards, monitoring, feedback, learning

support and knowledge representation) and the overall relationship between the pre-

scribed and the enacted curriculum—that our larger research program has addressed.

In the course of our analysis we will develop three arguments and intimate two

others that we have developed in our broader research program. The first is the sin-

gle-minded performative orientation of instructional practices generally—and

instructional strategies specifically—in Singaporean classrooms that rarely deviated

from a logic of curriculum coverage, knowledge transmission and reproduction

(assessment). We think this partly reflects the influence of underlying cultural

assumptions and institutional rules about education, teaching and learning—what

Jerome Bruner (1996) and David Cohen (1988) have separately termed a ‘folk peda-

gogy’—and the very considerable institutional authority of the national high stakes

assessment system in a society where the nexus between credentialing and social

mobility is unusually tight and the accountability system renders teachers unusually

susceptible to parent credentialing anxieties. Indeed, we think this deeply instrumen-

tal performative orientation largely explains the essential hybridity of instructional

practice in Singapore.

Second, while we found substantial evidence of a pervasive performative orienta-

tion to instruction, we also found that instructional practices did not conform to a par-

ticular theoretical or normative model of pedagogical practice. Teachers in Singapore

are not, by and large, sectarian or tribal in their instructional commitments—that is,

they do not tend to see themselves as members of a particular pedagogical sect or tribe

(traditional, constructivist) or, for that matter, East Asian or Western. Instead, they

Table 1. 2009 PISA results

Rank Country Overall reading Overall math Overall science

1 Shanghai, China 556 600 575

2 Korea 539 546 538

3 Finland 536 541 554

4 Hong Kong 533 555 549

5 Singapore 526 (5) 562 (2) 542 (4)

6 Canada 524 527 529

7 New Zealand 521 519 532

8 Japan 520 529 539

9 Australia 515 514 527

10 Netherlands 508 526 522

Source:OECD (2010).

58 D. Hogan et al.

© 2013 British Educational Research Association

Page 3: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

pragmatically mix and match, drawing from a variety of instructional perspectives in

ways that reflect their pragmatic, instrumental fit-for-purpose approach and broader

performative orientation. However, this non-sectarian pragmatism and hybridity is

neither culturally or institutionally innocent but reflects the play of powerful vernacu-

lar discourses (ability, readiness, meritocracy, nation-building, social harmony) and

the complex articulation of instructional practices in response to the national high

stakes assessment system and cross-cutting institutional and organizational processes

that variously support or constrain the alignment of instruction with assessment. On

balance though, the forces of alignment and institutional isomorphism have prevailed

decisively until now, although this is likely to slowly change in the years ahead as the

professionalization of teaching, might, if appropriately organized, offer at least some

protection of teaching and learning practices from these pressures.

Third, we found, like others before us, that the national high stakes assessment sys-

tem exercised considerable institutional authority over Singapore’s pedagogical sys-

tem through its unparalleled ability to shape the pattern of instructional practice at

the classroom level. In Singapore, the Ministry of Education administers three high

stakes national assessments: at the end of primary school (the Primary School Leaving

Examination [PSLE]) regulating access to secondary school and to the curriculum

streams within them, at the end of secondary school (the Cambridge O levels) regu-

lating access to post-secondary education, and at the end of Year 12 for those enrolled

in junior colleges hoping for admission to a university (the Cambridge A levels). The

national high stakes assessment system is intended to serve a number of objectives—to enhance the quality of teaching and learning, to provide a public, transparent, reli-

able and politically acceptable quality assurance and accountability mechanism, and

to provide a clearly meritocratic process for the allocation of students into school

streams and the social division of labour more broadly. This is a considerable politi-

cal, educational and ideological burden for any assessment system to bear. Not sur-

prisingly, it has long been the subject of close scrutiny by local commentators. Cheng

(1996, p. 9), for example, argued that the ‘examination is the soul of the ethos about

education in East Asian societies’ and ‘the goalkeeper to the quality of education out-

put’. Two years later, Cheah (1998, p. 4), similarly concluded that the teaching of

English language in Singapore was ‘driven by an “examination-type literacy”’ and

that teachers ‘teach in the way they believe will help more students pass their examin-

ations’. The trouble with this pedagogical arrangement, as Cheah goes on to con-

clude, is that the tight coupling or calibration of instruction and assessment has had

unfortunate and unintended consequences. He notes in particular that the introduc-

tion of a key pedagogical reform in English in the 1990s—process writing—was ‘ham-

pered by the emphasis on examinations in the school system’. More recently,

Anneliese Kramer-Dahl (2008), in her research on the implementation of the 2001

English syllabus in Singapore in 2006 and 2007, similarly reports that the English

teachers she and her colleagues studied closely repeatedly retreated from the innova-

tive pedagogy of the 2001 EL syllabus and fell back instead to the default position of

an examination-driven instructional regime in order that their students be properly

prepared for school-based and national high stakes assessments. ‘Indeed’, she writes,

‘“exams”, “practice” and “back to basics” were keywords reiterated over and over in

our interviews and in the classroom, often juxtaposed against what they see as an all-

Assessment and the logic of instructional practice 59

© 2013 British Educational Research Association

Page 4: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

too-unrealistic “advanced thinking and literacy” discourse of the syllabus’. One of her

two subjects, Elise, explained her teaching choices this way:

There’s the syllabus and there’s the exam. I feel a lot of the curriculum is controlled by the

exam. I’m building the kids up to a last final outcome, with lots of practice and drill…Ulti-

mately on paper, you have to have your exams and your tests. Making the link between

school and real life? I think for our students that’s not a salient incentive. You need to link

it to the exams, right? That’s the carrot that dangles. ‘You learn this because exams are

gonna have that’. For lots of our teenagers exams have become life. You can’t draw the

line, not in the Asian context, because it’s very very obvious, you know. (Kramer-Dahl,

2008, p. 94)

Still, it is fair to say that Singapore’s national high stakes regime has had both posi-

tive and negative consequences for the quality of teaching and learning. On the one

hand, it goes a long way towards explaining the clear-eyed focus, coherence and effec-

tiveness of instructional practice and the underlying performative pedagogical orien-

tation that underwrites it in Singapore. On the other hand, we also think that the

national high stakes assessment system has resulted in a pedagogy that is intractably

didactic rather than dialogical, compromised the epistemic quality and the transpar-

ency or ‘visibility’ (Hattie, 2009, 2012) of learning processes during lessons,

restricted the opportunities of students to engage in knowledge building work in class,

and constrained the ability of the system to successfully introduce substantial and sus-

tainable pedagogical improvements despite a strong policy commitment to doing so

as reflected in the two key policy documents of the past 15 years—Thinking schools,

learning nation (TSLN, 1997) and Teach less, learn more (TLLM, 2004). We consider

these very high opportunity costs to pay and to reflect a considerable lack of align-

ment between policy and practice in the system, even though at the same time they

reflect a very high degree of pedagogical alignment between assessment and instruc-

tion. Indeed, our findings suggest a very considerable tension, if not outright contra-

diction, between the teaching for understanding and twenty-first century learning

objectives of recent policy statements (especially Teach less, learn more) and the

continuing commitment of the government to its national high stakes assessment

regime. When all is said and done, these findings raise important questions about

whether the current pedagogical model, originally developed in the late 1970s, has

now more or less run its course and needs substantial modification of its basic design

principles if the system is to have any real hope of achieving the policy priorities set

out in TSLN and TLLM. Tinkering around the edges—a little bit more feedback and

formative assessment here, a little bit more teaching for understading there, a little

more PD everywhere—is unlikely to achieve the outcomes the system desires.

In short, over the course of our article we develop three key arguments—one regard-

ing the overall performative orientation of instructional practice in Singapore, the sec-

ond regarding the pragmatic hybridity of instructional practice, and the third the tight

coupling or alignment of national high stakes assessment system and classroom

instruction. We think these arguments go some way towards explaining why Singa-

pore’s educational system has achieved extraordinary success in a very short period of

time, accelerating from a standing start in 1965 as a third world education system to

one of themost successful 40 years later.We think this remarkable achievement can be

explained, in part, by the single-minded instrumentalist focus of the system on devel-

60 D. Hogan et al.

© 2013 British Educational Research Association

Page 5: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

oping a highly effective performative pedagogy using a high stakes national assessment

system to lift standards and ensure quality. Of course, there are other factors as well

that have contributed to Singapore’s success, including a high quality curriculum

framework, exceptional leadership of a tightly integrated and centralized system, the

quality and commitment of its teaching corps, and highly supportive cultural orienta-

tions to education that in part reflects Singapore’s Confucian heritage and in part the

tight nexus between educational credentialing and status attainment in Singapore. But

our evidence also suggests that the basic design principles of the pedagogical system

that have served Singapore sowell over the course of the developmental phase of its his-

tory now appear, ironically enough, to constrain the ability of the system to successfully

support substantial and sustainable pedagogical improvement in line with its own pol-

icy preferences. This has important implications for other systems that have recently

made it their business to emulate or imitate some key features of Singapore’s system.

Data andmethods

The data we report in this study draws from a nationally representative sample of over

4000 Secondary 3 students and their teachers in approximately 120 mathematics and

English classes across 32 secondary schools in Singapore conducted in 2010. Our first

interim report (Hogan et al., 2011) includes detailed discussions of the statistical proce-

dures specification and theoretical issues involved in the construction of all of the

instructional methods scales. However, it is appropriate now to briefly discuss some

details of our sampling strategy and the reliability of students as raters. In order to

maximize the analytical scope of key pedagogical practices and individual student

characteristics, we employed a split-half multi-level strategy in which 50% of the total

student samples within each class in the sample were randomly assigned to a 230-item

survey focused on students’ perceptions of instructional practices. The other half of

the samples were assigned to a survey in which students answered a similar number of

questions about their family background, learning orientation, motivation, self-beliefs

and so on. In addition, all the students in each class sampled completed an hour-long

assessment in mathematics or English. This design permits us to model the logic of

instructional practice at the classroom level and to model the impact of domain spe-

cific classroom practices on student outcomes. In this article, however, we draw

exclusively on the classroom level survey data only, apart from limited interview data

with teachers concerning their views on assessment.

While relatively few studies have been conducted of primary and secondary school

students as raters of instructional practices, the studies that have been completed

indicate that elementary and secondary school students can effectively differentiate

among various types of instructional practices (Urdan & Midgley, 2001; Wolters,

2004; Assor et al., 2005). In any case, there are compelling statistical reasons to use

the multiple data points from student reports at the class level to model instructional

practices rather than the single data point provided by the class teacher.

While we report a range of descriptive statistics in the analysis that follows, analyti-

cally the greater part of the heavy lifting is undertaken by structural equation model-

ling (SEM). SEM is a powerful multivariate statistical technique used to examine

whether the hypothetical relationships specified by the researcher can be empirically

Assessment and the logic of instructional practice 61

© 2013 British Educational Research Association

Page 6: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

supported by the sampled data. One of its key strengths is its ability to specify latent

variables as robust measures of theoretical constructs. Unlike traditional factor ana-

lytic approaches where theoretical variables of interests are often assumed to be mea-

sured without error, measures assessed using confirmatory factor analysis (the

measurement approach of SEM) are corrected for biases attributable to random and

unexplained measurement error that often result in more precise estimates of struc-

tural relationships (Bollen, 1989). Another strength of SEM is the availability of good-

ness-of-fit statistics that allow researchers to evaluate the relative strength of the

hypothesized model against the baseline structure. In this study, evaluation of good-

ness-of-fit statistics is based on the classical chi-square test (v2) where a non-signifi-

cant v2 at a specified alpha level of p < .05 indicates that there is no statistically

significant discrepancy in the covariance structure between the observed data and the

hypothesized model. However, since the v2 can be sensitive to sample size and model

complexity, we also report a range of alternative fit statistics together with the v2 testthat take these issues into account. Based on the conventional guidelines by Hu and

Bentler (1999), the recommended thresholds of a good fitting model are as follows:

comparative fit index (CFI) and Tucker–Lewis index (TLI) should have values above

0.95, root mean square error of approximation (RMSEA) should have values less than

0.05, and standardized root mean square residual (SRMR) should have values less

than 0.08.

Although researchers would typically like to develop empirical models that are suf-

ficiently complex so that they are closer to explaining the truth as observed in the real

world, this is often challenging as complex models can result in unstable parameter

estimates (Bandalos & Finney, 2001). One strategy often employed (and the one

employed in this article) is the use of composite variables or ‘item parcels’ as proxies

for the estimated latent variable. As composite variables reduce the number of param-

eters that must be estimated by the model, complex analyses requiring a larger pool of

variables become less demanding. This, happily, reduces sampling and measurement

error, thus resulting in more accurate estimates. In this study, a single-item latent

composite variable is constructed by assigning a pre-calculated regression and error

constraint estimated from a proportionally weighted composite score. Given that uni-

dimensionality is an important condition for latent composite models (Sass & Smith,

2006), all instructional variables presented in this article are psychometrically reli-

able. Construct reliability ranged from .79 to .90, and all over-identified latent mea-

surement models achieved a non-significant chi-square (p < .05), except for one

(p = .001). The general approach to this procedure is explained in more detail by

Chan (2012). Similar approaches to this procedure has been established previously as

total aggregation method with reliability correction (Coffman & MacCullum, 2005),

equivalent composite measures (Landis et al., 2000), latent composite variables (Ste-

phenson & Holbert, 2003), one-factor congeneric measurement model (J€oreskog,1971) and one-factor congeneric measurement model (Rowe, 2002).

The analysis of instructional methods and student achievement reported in this

article is part of a larger three year study (the Core 2 Research Program) focused on

mapping, measuring and modelling instructional practices in Singaporean classrooms

and their impact on wide variety of student outcomes (Hogan et al., 2011). We have

already published a number of papers that focus on instructional tasks, classroom talk

62 D. Hogan et al.

© 2013 British Educational Research Association

Page 7: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

and student motivation (Hogan et al., 2011, 2012a, b; Luo et al., 2011a, b; Rahim et

al., 2012) and will not address these issues here. This particular article focuses, as the

title suggests, on the logic of instructional strategies where we understand these to be

composed of sets or arrays of instructional methods loosely organized as very general

instructional strategies—‘traditional instruction’, ‘direct instruction’, ‘teaching for

understanding’ and ‘co-regulated learning strategies’. We recognize that only the

more theoretically self-conscious of teachers will frame their teaching in these broad

strategic terms (let alone even broader theoretical frameworks), and that they are far

more likely to understand their practice in somewhat more disaggregated and discrete

terms. Our theoretical models then function as methodological heuristics that allow

us to explore the relationships within and between nominally independent sets of

instructional strategies. Intriguingly, however, statistically speaking, the clusters of

discrete instructional practices associated with particular strategies exhibit relatively

high levels of covariance and therefore constitute relatively coherent clusters of

instructional practice that we might well consider empirically established latent theo-

retical models of instructional practice. However, we hasten to add that the covari-

ances between the sets of instructional practice are not so low that teachers, at least in

Singapore, feel compelled to choose between them rather than combine them in ways

that they believe are pragmatically ‘fit for purpose’. We also recognize the theoretical

risks in treating instructional strategies separate from instructional tasks, lesson

organization, curriculum materials, classroom management, learning environment

and talk but treating all of them simultaneously in one publication places impossible

demands on editors of academic journals.

A taxonomy of instructional methods

Traditional instruction. One common stereotype of East Asian pedagogy is that it is

characterized by ‘traditional’ forms of instruction and that this is a major part of the

explanation of why East Asian students have done so well in international assess-

ments like TIMMS and PISA. In its cruder forms the East Asian thesis is that drill

and practice supports memorization and memorization develops understanding;

more sophisticated versions suggest that drill and practice develops both procedural

skills and conceptual understanding as well, although this version of the thesis is

unclear about the nature of the causal relationships involved. Although by no means

the first to do so, Frederick Leung describes the stereotypical image (particularly in

the West) of the teaching style in mathematics in East Asia in the following terms.

Mathematical classes, he writes, are seen as ‘rather traditional’. Teaching is ‘predom-

inantly content orientated and exam driven. Instruction is very much teacher domi-

nated and student involvement minimal’. Teaching is ‘usually conducted in whole

group settings, with relatively large class sizes’. There is ‘virtually no group work or

activities, and memorization of mathematics is stressed’ and ‘students are required to

learn by rote’. Students are ‘required to engage in ample practice of mathematical

skills, mostly without thorough understanding’ (2001, pp. 35–36; see also Biggs &

Watkins, 2001; Cai et al., 2004; Wong, 2004; Leung, 2006). Leaving aside whether

this is an accurate picture in East Asia, is it an accurate guide to teaching in

Singapore?

Assessment and the logic of instructional practice 63

© 2013 British Educational Research Association

Page 8: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

Our particular specification of what we call ‘Traditional Instruction’ (TI) focuses

on five first order constructs or single indicators: a focus on worksheets and work-

books (‘How often does your mathematics/English teacher ask you to do worksheets

or workbooks?’); a focus on textbooks (e.g., ‘How often does your mathematics tea-

cher asks you to answer questions from the textbook?’); drill and practice of basic

facts, rules and procedures (e.g., ‘How often does your mathematics/English teacher

ask you to drill and practice on basic facts, rules or procedures?’); a focus on memori-

zation (e.g., ‘How often does your mathematics teacher ask you to remember formu-

lae or rules?’); and exam preparation (‘my teacher emphasizes studying problems that

may occur in the exams’; ‘my teacher spends a lot of class time preparing for exams’;

‘my teacher teaches us test-taking strategies’; and ‘my teacher emphasizes practicing

past year exam papers’). Initially, we also included an item ‘teacher talks/lectures a

lot’ on the grounds that theoretically at least it ought to fit the TI model, but when we

subjected the model to further statistical analysis using confirmatory factor analysis

(CFA) to test the reliability of the TI scale, teacher talk/lecture did not load well on

the TI second order construct in either English or mathematics. Accordingly, we did

not include it in our final model. Some of the measures in the TI scale are made up of

single indicators measures: we recognize that this is far from optimal statistically (we

would have preferred three or four), but the post hoc nature of our analysis of the TI

scale (and some other measures) gave us little choice.

Table 2 reports the results of the CFA models of mathematics and English. Factor

loadings range from the moderately high (.790, .841) to, in the case of English, the

relatively low (.426). The CFA models indicate quite different factor structures for

the two subjects. In mathematics, textbook focus has the strongest loading, followed

by memorization and drill and practice. By contrast, in English, textbook focus has

the lowest factor loading of all; conversely, memorization has the strongest factor

loading, followed by drill and practice and exam preparation. The re-specified model

includes a theoretically sensible correction for an error covariance in both models:

Table 2. CFA higher order factor loadings and goodness-of-fit statistics: traditional instruction,

Secondary 3 mathematics and English

Mathematics English

Respecified higher order model Respecified higher order model

Higher order factor loadings

Textbooks focus (2) .790 (.035) .426 (.034)

Memorization (1) .784 (.023) .841 (.022)

Drill and practice (1) .771 (.023) .800 (.022)

Worksheets focus (1) .648 (.027) .573 (.030)

Exam preparation (4) .611 (.027) .784 (.022)

Goodness-of-fit statistics:

Chi-square/df/p-value 48.115/24/.0024 30.143/18/.0361

CFI/TLI .993/.989 .996/.993

RMSEA (90% C.I.) .029 (.017–.041) .026 (.007–.041)SRMR .021 .016

64 D. Hogan et al.

© 2013 British Educational Research Association

Page 9: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

between exam preparation and drill and practice in mathematics, and between text-

book focus and drill and practice in English.

The mean scores, standard deviations and Cronbach’s alphas for the scales are

reported in Table 3. The overall mean scores for traditional instruction in both Eng-

lish and mathematics are relatively high, although there are important differences

across subjects—mathematics teachers especially are far more likely to employ tradi-

tional instructional techniques in general than English teachers. However, the rank

ordering across the two subjects differs slightly: in mathematics, memorization, fol-

lowed by focus on worksheets and workbooks and focus on textbooks, are dominant;

in English, focus on worksheet sand workbooks, memorization and focus on exam

preparation led the field.

The mean scores tell us about the relative frequency of instructional events; what

they cannot tell us about are the strength of the relationships between the five mea-

sures. CFA modelling of course provides a measure of patterns of covariance; the

CFAs reported in Table 1 indicate reasonably strong levels of covariance. A closely

related statistic—correlation coefficients—provides a second. These are reported in

Table 4 and again indicate moderately strong relationships, although the strength of

the relationships varies between subjects, but with particularly strong correlations in

both subjects between exam preparation and focus on drill and practice.

Table 4. Latent correlation matrix: traditional instruction for mathematics and English

(mathematics to the left of the diagonal; English to the right)

EXPR MEM DRILL WKSF TXBF

Focus on exam preparation .65 .62 .45 .37

Focus on memorization .45 .68 .50 .32

Focus on drill and practice of basic

facts, rules and procedures

.53 .59 .44 .35

Focus on worksheets and workbooks .39 .51 .50 .40

Focus on textbooks .44 .67 .58 .51

Note: All correlations significant at p < .01.

Table 3. Traditional instruction, Secondary 3 mathematics and English

Mathematics English

Mean (1–5) SD Mean (1–5) SD Cohen’s d

N 1166 1027

Traditional Instruction (TI)

Scale (alpha: .758, .726)

3.78 .65 3.46 .65 .49

Focus on memorization 4.06 .91 3.47 .93 .64

Focus on worksheets and

workbooks

3.93 .95 3.63 .95 .32

Focus on textbooks 3.83 .82 3.35 .97 .54

Focus on drill and practice on

basic facts, rules and procedures

3.59 .97 3.39 .92 .21

Focus on exam preparation 3.51 .86 3.45 .82 .07

Assessment and the logic of instructional practice 65

© 2013 British Educational Research Association

Page 10: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

Neither CFA nor bivariate correlations, however, tell us much about the pattern of

causal relationships between variables. Is it the case, for example, as pedagogical folk-

lore in Singapore suggests, that instructional practices in Singapore are more or less

dominated by the institutional authority of the assessment system? To even begin to

answer this question we need a kind of statistical analysis that can estimate the

strength of ‘causal’ pathways between variables controlling for the influence of con-

founding factors. The statistical procedure of choice these days to answer this kind of

question is structural equation modeling (SEM). SEM models allow researchers to

precisely measure the pattern and strength of relationships between the different

instructional practices (technically, ‘constructs’) by estimating regression coefficients

for the pathways between the constructs. These regression coefficients predict how

much of the variance in the outcome variable (the construct into which the arrow

goes) is accounted for by variations in the predictor variable (the construct from

which the construct leaves) controlling for the influence of other variables on the

outcome variable.

SEM models proceed by first specifying a theoretical model of the relationships

that researchers hypothesize exists in the real world, and then tests to see how well the

real world empirical data ‘fits’ the model. The better the fit, the better the model. In

addition, SEM models estimate the strength of relationships (pathways) between

variables and report the results in terms of regression coefficients. A pathway regres-

sion coefficient of .47, for example, between exam preparation and textbook focus

means that for every unit increase in exam preparation, teachers will increase their

focus on textbooks by .47 of whatever unit metric is being used, as in the SEMmodel

for Secondary 3 mathematics below (Figure 1). This means that a focus on exam

preparation is directly predictive of textbook focus (with a strong regression coeffi-

cient of .47). In addition, exam preparation is predictive of a focus on worksheets and

workbooks (.18), and the drill and practice of basic facts, rules and procedures (.34).

We believe that the strength of these coefficients underscores the institutional ability

of the assessment system, through its classroom proxy (exam preparation), to shape

Note: Values represent unstandardized estimates significant at p < .01

Goodness-of-fit statistics:Chi-square/df/p-value 1.902/1/.1679CFI / TLI .999/.993RMSEA (90% CI) .028 (.000–.089)SRMR .006

TextbookFocus

DrillExam

Preparation

Worksheet Focus

.34(.04) .25(.05)

.31(.05)

.21(.04)

Memorization

.18(.04)

.47(.04)

.15(.04).43(.05)

.46(.06)

Figure 1. SEMmodel for traditional instruction (mathematics)

66 D. Hogan et al.

© 2013 British Educational Research Association

Page 11: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

classroom practice. But, in addition, all three of these strategies in turn have direct

pathways to a focus on memorization, with a particularly strong pathway from text-

book focus to memorization (.46). Further, there are a series of indirect pathways

from exam preparation to memorization. All together these indirect pathways add up

to a substantial effect size of .42 if we multiply out and add up all the individual coeffi-

cients ((.47*.46) + (.34*.25) + (.18*.15) + (.18*.21*.25) + (.47*.31*.25) + (.47.

*43*.15) + (.47*.43*.21*.25)). Effect sizes of this kind should be interpreted in much

the same way as Cohen’s d; in this case, an effect size of .42 is, by convention, consid-

ered a small to moderate effect size. As a standard SEM convention, unstandardized

estimates are reported throughout this article (see Brown, 2006; Kline, 2011).

Although standardized coefficients are often used to infer the magnitude of parameter

estimates, there is no reason why unstandardized estimates cannot be interpreted as

effect sizes since all measures adopted a similar response metric (i.e., 1 to 5) (see

Preacher & Kelley, 2011). Moreover, as single-item latent variable models were used,

comparisons between standardized estimates found no appreciable differences, and

even if differences were observed, they were often found at the second or third deci-

mal point.

In Secondary 3 English (Figure 2), model fit is also very good (although not quite

as good as in mathematics). Exam preparation, similarly, has a pervasive and

substantial impact on all other traditional instruction practices: textbook focus (.38),

a focus on worksheets and workbooks (.35), drill and practice of basic facts, rules and

procedures (.54), and memorization (.32). Exam preparation, drill and worksheet

focus in turn have direct pathways to a focus on memorization (.32, .41 and .17,

respectively). In addition, there are a series of indirect pathways from exam prepara-

tion to memorization via textbook focus, worksheet focus and drill. All together these

direct and indirect pathways from exam preparation to memorization add up to a very

large effect size of .66 if we multiply out and add up all the coefficients (.32 +(.38*.28*.17) + (.38*.28*.20*.41) + (.54*.41) + (.35*.17) + (.35*.20*.41)). This

result underscores the very substantial leverage exam preparation has over TI prac-

tices more generally.

Note: Values represent unstandardized estimates significant at p < .01

Goodness-of-fit statistics:(N = 1027)Chi-square/df/p-value 5.569/2/.0617CFI/TLI .997/.985RMSEA (90% CI)/SRMR .042 (.000–.085)/.012

TextbookFocus

DrillExam

Prepara on

Worksheet Focus

.54(.04) .41(.04)

.20(.04)

Memoriza on

.35(.04)

.38(.04)

.17(.04).28(.04)

.32(.04)

Figure 2. SEMmodel for traditional instruction (English)

Assessment and the logic of instructional practice 67

© 2013 British Educational Research Association

Page 12: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

Direct instruction. Direct Instruction (DI) has a well-established and well-deserved

track record for its ability to promote student learning for a broad variety of learning

tasks, including the acquisition of content knowledge and procedural skills, and the

execution of relatively straight forward cognitive tasks (Desforges, 1995; Good &

Brophy, 2003, Chapter 9; Hattie, 2003, 2009; Rowe, 2003, 2006; Purdie & Ellis,

2005; NMAP, 2008, Chapter 4). Hattie (2009), however, claims the efficacy of direct

instruction across a broader range of tasks (with a Cohen’s d effect size of .59),

although he also employs a relatively broad conception of direct instruction that is

conceptually close to related notions of mastery learning and active teaching. Our

specification of DI is narrower and includes five first order constructs: a four-indica-

tor measure of maximum learning time (e.g., ‘The teacher makes sure that pupils

focus on the lesson’); a four-indicator measure of review (e.g., ‘The teacher checks

that pupils understand the lesson’); a six-indicator measure of structure and clarity

(e.g., ‘The teacher clearly states the objectives of the lesson’, ‘The teacher organizes

information in an orderly way’, ‘The teacher explains things very clearly’); a single

indicator measure of time on practice (‘We spend a lot of time practicing what we

learned’); and a one-indicator measure of frequency of questioning (‘The teacher asks

the class lots of questions’). The last two scales, having only one indicator each, are

rather thin and far from what we would like to have, but our modelling of the con-

structs indicates that they behave sensibly and usefully, statistically speaking.

The results of the confirmatory factor analysis, reported in summary form in

Table 5, indicate good fit statistics in both subjects. Factor loadings in both subjects

range from the high (.918, .958 respectively in mathematics and English) to the mod-

erately low (.587, .586). There was only one significant error covariance in either sub-

ject between the sole frequency of questioning indicator (‘the teacher asks the class

lots of questions’) and the sole indicator of the frequency of practice.

The mean scores, standard deviations, alphas and effect sizes for the DI scales are

reported in Table 6. In both mathematics and English, maximum learning time

Table 5. CFA higher order factor loadings and goodness-of-fit statistics: direct instruction,

Secondary 3 mathematics and English

Mathematics English

Respecified higher order model Respecified higher order model

Higher-order factor loadings:

Revision (4) .918 (.015) .958 (.015)

Structure and clarity (6) .849 (.015) .838 (.016)

Maximum learning time (4) .828 (.016) .859 (.015)

Time on practice (1) .677 (.023) .689 (.024)

Frequency of questioning (1) .587 (.026) .586 (.027)

Goodness-of-fit statistics:

Chi-square/df/p-value 127.377/61/.0000 166.298/61/.0000

CFI/TLI .991/.989 .985/.981

RMSEA (90% C.I.) .031 (.023–.038) .041 (.034–.049)SRMR .020 .024

68 D. Hogan et al.

© 2013 British Educational Research Association

Page 13: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

scores the highest mean values, followed by structure and clarity and review. These

are important findings that highlight key strengths of Singaporean pedagogy and play

important roles in shaping the overall structure of instructional methods.

As we did before with Traditional Instruction, we can also examine the strength of

the relationships between the DI measures by looking at correlation coefficients

(Table 7). Again, the coefficients are sensible and slightly higher in general in English

than they are for mathematics. However, the absolute values of the coefficients are

generally higher than they are in TI.

But, as with TI, correlation coefficients do not help us understand the strength of

the relationships between the practices controlling for the confounding influence of

other practices. Again, we can represent the strength of these relationships with struc-

tural equation models (Figures 3 and 4). The DI SEM model for both English and

mathematics fit the data exceptionally well, although the model fit is not quite as

strong in English as it is for mathematics. Still, the internal pathway structure of the

two broad instructional strategies are very similar, although not identical, in that DI

for English includes a non-recursive relationship between teacher revision and fre-

quency of practice missing in mathematics. All the structural relationships repre-

sented by the model are strong and statistically significant. On balance, coefficients

are slightly higher in English. In our construction and interpretation of both the math-

ematics and English models we drew upon conventional conceptions of direct

instruction and John Hattie’s (2009) account of visible teaching and learning to

Table 6. Direct instruction, Secondary 3 mathematics and English

Mathematics English

dMean (1–5) SD Mean (1–5) SD

N 1166 1027

Direct instruction

(alpha: .844, .850)

3.61 .668 3.53 .65 .12

Maximum learning time (4) 3.89 .767 3.84 .78 .06

Structure and clarity (6) 3.61 .812 3.56 .78 .06

Teacher revision (4) 3.59 .835 3.52 .77 .09

Frequency of practice (1) 3.49 .952 3.30 .94 .20

Frequency of questioning (1) 3.47 .946 3.44 .91 .03

Table 7. Latent correlation matrix: direct instruction for mathematics and English (mathematics

to the left of the diagonal; English to the right)

MLT REV SNC FOQ FOP

Maximum learning time .85 .72 .48 .53

Teacher revision .78 .78 .55 .67

Structure and clarity .70 .77 .53 .64

Frequency of questioning .47 .55 .51 .58

Frequency of practice .53 .60 .62 .61

Note: All correlations significant at p < .01.

Assessment and the logic of instructional practice 69

© 2013 British Educational Research Association

Page 14: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

hypothesize that structure and clarity would perform a similar structuring or organizing

role in direct instruction as exam preparation does in traditional instruction. Accord-

ingly, we have treated it as an exogenous variable in both models.

In mathematics, for example, there are strong pathways from structure and clarity

to maximum learning time (.70), indicating that teachers who have structure and

clarity in their lessons are very likely to maximize learning time in their classes. In

English the pathway is slightly stronger (.72). Structure and clarity also has a strong

Note: Values represent unstandardized estimates significant at p < .01

Goodness-of-fit statistics:(N = 1166)Chi-square/df/p-value 2.296/3/.5134CFI/TLI 1.00/1.00RMSEA (90% CI) .000 (.000–.04)SRMR .006

Teacher Revision

Structure & Clarity

MaximumLearning

Time

Frequency of Ques oning

.70(.03) .43(.04) .30(.04)

.48(.04)

Frequency of Prac ce

.31(.06)

.43(.04).39(.06)

Figure 3. SEMmodel for direct instruction (mathematics)

Note: Values represent unstandardized estimates significant at p < .01

Goodness-of-fit statistics:(N = 1027)Chi-square/df/p-value 4.421/2/.1097CFI/TLI .999/.994RMSEA (90% CI) .034 (.000–.079)SRMR .008

TeacherRevision

Structure &Clarity

MaximumLearning T

Frequency ofPrac ce

.72(.03)

.27(.05)

.33(.05)

.58(.04)

Frequencyof

Ques oning

.14(.05)

.37(.05)

.47(.07)

.23(.08)

Figure 4. SEMmodel for direct instruction (English)

70 D. Hogan et al.

© 2013 British Educational Research Association

Page 15: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

direct pathway to teacher revision in mathematics (.43) (a measure of the extent to

which teachers revise the work they have done with students previously), and a some-

what weaker pathway to teacher revision in English (.27), a finding consistent with

the conventional folk wisdom about the relative importance of revision in mathemat-

ics. The pathway from structure and clarity in mathematics to frequency of practice is

also strong (.39), and even stronger in English (.47), which probably reflects the

greater attention in English to practicing genre-based work rather than drilling. All

three of these pathways are theoretically and practically sensible in both subjects. In

mathematics both teacher revision (.30) and frequency of practice (.43), in turn, have

direct pathways to frequency of questioning, our key outcome measure in the model.

In English, the two pathways to frequency of questioning have coefficients of .33 (tea-

cher revision) and .37 (frequency of practice) respectively.

Intriguingly, there are no statistically significant direct pathways from structure and

clarity to frequency of questioning in either mathematics or English. Instead, the

effect of structure and clarity on frequency of questioning is indirect in both subjects,

with a total, and very substantial, effect size of .495 in mathematics and .490 in

English. This is an important finding, underscoring the complementary role that tea-

cher questioning can play in enhancing the structure and clarity, and hence the visibil-

ity of teaching and learning, to both teachers and students. In English there is also a

non-recursive relationship between teacher revision and frequency of practice, with a

coefficient of .23 for the pathway from teacher revision to frequency of practice, and a

weaker but still significant pathway from frequency of practice to teacher revision

(.14). This too makes pedagogical sense, in that it suggests that not only do teachers

use revision as a springboard to practice, as we would expect teachers to do, but they

also respond to practice with additional revision if they feel that the students do not

quite understanding what they need to learn. In mathematics the relationship is one

way, from teacher revision to frequency of practice (.30), suggesting in this matter at

least, that mathematics teachers could learn a thing or two from their English

colleagues.

Note: Values represent unstandardized es mates significant at p < .01

Goodness-of-fit sta s cs:(N = 1166)Chi-Square / df / p-value 38.300 / 19 / .0054CFI / TLI .996 / .990RMSEA (90% CI)/ SRMR .030 (.016-.043) / .012

ExamPrepara on

TeacherRevision

Structure &Clarity

MaximumLearning Time

Frequency ofQues oning

.34(.05)

.46(.04)

.10(.04)

.17(.05)

.23(.04)

.13(.05)

.19(.05).28(.05)

.14(.05)

.16(.05)

Frequency ofPrac ce

TextbookFocus

Drill

WorksheetFocus

Memoriza on

.25(.06)

.48(.04)

.18(.04)

.43(.05)

.34(.04)

.33(.05)

.25(.05)

.16(.04)

.45(.06)

.20(.04)

.44(.04)

.31(.05)

.17(.06)

.56(.03)

.29(.04)

.13(.04)

Figure 5. SEMmodel for traditional and direct instruction (mathematics)

Assessment and the logic of instructional practice 71

© 2013 British Educational Research Association

Page 16: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

The next step in our analysis was to examine the relationships between the TI and

the DI models. We first conducted an integrated CFA of the two sets of measures

followed by an integrated SEM model that incorporated both TI and DI practices

(Figures 5 and 6). We only report the SEM results here. A number of features of the

integrated SEMmodels should be highlighted.

First, the internal structure of the two sets of instructional strategies remained

remarkably stable in the models, with almost identical networks of pathways in the

integrated models that we achieved separately and discussed above. Second, there is

significant asymmetrical interaction between the TI and DI instructional strategies.

Indeed, in mathematics, all the TI constructs have pathways leading to DI constructs,

but none in the reverse direction. This is true as well of English with but one excep-

tion. We tested models that reversed the relationship between DI and TI, as well as

looked for non-recursive relationships (that is, feedback loops) from DI to TI con-

structs. In mathematics this was to no avail. Instead we found that our best fitting

models were ones that ran the pathways from TI to DI. In English, however, we did

identify one pathway from a DI practice to a TI practice: from revision to textbook

focus (.25), indicating that teachers often, at least in English, follow up a revision ses-

sion with a textbook session.

Third, we concluded that the asymmetrical relationship between TI and DI is prin-

cipally a function of the institutional leverage exam preparation has over instructional

practices in Singapore in both subjects. Indeed, there are multiple direct and indirect

paths from exam preparation to other traditional instructional methods and to direct

instructional methods. Because we have positioned it (after testing a number of alter-

native models) as an exogenous variable in the model, it has no pathways into it. But

in mathematics it has a total of seven pathways leading from it towards other TI vari-

ables and all but one of the DI measures as well (maximum learning time). Three of

the coefficients are quite strong—textbook focus (.48), structure and clarity (.46) and

drill (.34)—but most are small to modest. In English, there are six direct pathways

Note: Values represent unstandardized estimates significant at p < .01

Goodness-of-fit statistics:(N = 1027)Chi-square/df/p-value 47.613/20/.0005CFI/TLI .994/.986RMSEA (90% CI) .037 (.023–.050)SRMR .015

ExamPreparation

TeacherRevision

Structure &Clarity

MaximumLearning

Time

Frequency ofQuestioning

.26(.05)

.52(.04)

.16(.04)

.14(.04)

.34(.04)

.19(.06)

.13(.03)

.17(.05)

.24(.05)

Frequency ofPractice

TextbookFocus

Drill

WorksheetFocus

Memorization

.22(.07).21(.06)

.37(.04).25(.04)

.54(.04)

.42(.04).14(.04)

.18(.04)

.49(.04)

.25(.05).23(.08)

.67(.04)

.30(.04)

.27(.06)

.20(.03)

.15(.06)

Figure 6. SEMmodel for traditional and direct instruction (English)

72 D. Hogan et al.

© 2013 British Educational Research Association

Page 17: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

from exam preparation to TI and DI practices. However, in English, there are no

pathways from exam preparation to teacher revision or frequency of questioning,

indicating that the direct institutional leverage of the examination system over DI

instructional practices in English is slightly lower than it is in mathematics. But for

both subjects, all of the pathways are theoretically sensible, and attest to the institu-

tional authority of the assessment regime over instructional practices in a way that

strengthens the alignment—or, to change the metaphor, tightens the coupling—of

classroom instruction and the national assessment regime.

Fourth, in integrating the two instructional strategies, we positioned the DI out-

come variable, frequency of questioning, as the outcome variable for the integrated

model for both subjects. We did so for three reasons: it worked well statistically in our

SEM model of DI, the importance of teacher questioning and classroom talk more

generally as a gateway to the co-construction of meaning and conceptual understand-

ing, and the fact that we were unable to get as good a fitting model with memorization

as the outcome measure. This is not to say that teachers did not value or employ

memorization as an instructional strategy—but they employed it rather differently in

the two subjects. In mathematics, as we know from Table 3, memorization had the

highest mean score of all the TI instructional practices (4.06), indicating that teachers

relied on it a lot to promote a particular kind of learning that we can assume reflects

their preoccupation with exam preparation. Indeed, it is notable that memorization is

causally dependent on three other instructional methods within TI ensemble: from

worksheet focus (.16), from textbook focus (. 45) and from drill and practice of basic

facts, rules and procedures (.25). But while mathematical teachers clearly find memo-

rization useful as a follow up to other traditional instructional methods, they do not

find it particularly useful as a platform to scaffold further instructional strategies.

Rather, they are more likely to use it as a means of completing a teaching–learningsequence than begin a new one. In English, on the other hand, the mean score (3.47)

indicates that English teachers rely relatively less on memorization than mathematics

teachers do, but they also view it as more potent, instructionally generative and

instrumentally useful than their mathematics colleagues.

Fifth, overall, the total effect size, including both direct and indirect pathways from

exam preparation to frequency of questioning, our key outcome practice, for mathe-

matics is .53 (.17 + .36). For English, it is .48 (indirect effect only as the direct path

was not statistically significant).

Sixth, in Table 3 we reported that focus on textbooks has a relatively high mean

score (3.83) in Secondary 3 mathematics and a relatively low score in English (3.35).

In the SEMmodel for mathematics, textbook focus has only one pathway leading into

it—from exam preparation—but it is highly generative in its impact on other instruc-

tional practices, with seven pathways leading from it to other instructional practices.

Three of these are to other TI practices (drill (.33), memorization (.45) and work-

sheet focus (.43). All of these coefficients are substantial and theoretically meaning-

ful, indicating that the use of textbooks is a prelude to (and therefore predicts) the

other three TI practices. Four other pathways from textbook focus in mathematics

lead to four DI practices: structure and clarity (.13), maximum learning time (.19),

frequency of practice (.28) and frequency of questioning (.13). Overall, the total

effect size of the pathways from textbook focus to frequency of questioning is .17. In

Assessment and the logic of instructional practice 73

© 2013 British Educational Research Association

Page 18: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

English, the textbook focus is much less strong and pervasive in its impact. Like

mathematics, it only has one pathway into it, from exam preparation (.21), but unlike

mathematics, it only has three (rather than seven) pathways leading from it: to work-

sheet focus (.25), frequency of practice (.13) and frequency of questioning (.14). This

suggests that English teachers rely less on textbooks to shape their instruction than

teachers do in mathematics. However, the SEM model indicates that there is moder-

ately strong feedback loop from teacher revision (.27) back into textbook focus, indi-

cating that teachers often follow up a revision session with a return to the textbook in

use, and so the cycle begins again.

Seventh, two DI practices play structurally important roles generatively: structure

and clarity, and teacher revision. In mathematics, structure and clarity is predicted by

three TI practices, two of them with substantial coefficients: textbook focus (.13),

drill (.29) and exam preparation (.46). This is an important finding, attesting to

the causal importance of TI, particularly exam preparation, given the size of its

coefficient, to the enactment of DI practices. In English, structure and clarity is pre-

dicted by three TI practices: drill (.30) and exam preparation (.52). No DI practices

predicted structure and clarity in either mathematics or English. In turn, as a predic-

tor variable, in both mathematics and English, structure and clarity has three path-

ways from it to other DI practices: maximum learning time (.56 and .67 respectively),

teacher revision (.31 and .25 respectively) and frequency of practice (.25 and .22

respectively). There is no statistically significant pathway from structure and clarity to

frequency of questioning in either mathematics or English, replicating our finding of

the SEM model of DI above. But structure and clarity does have an indirect effect on

frequency of questioning through four pathways in mathematics and English—from

structure and clarity to maximum learning time (MLT) to teacher revision to fre-

quency of questioning; from structure and clarity to teacher revision to frequency of

questioning; from structure and clarity to teacher revision to frequency of practice to

frequency of questioning; and from structure and clarity to MLT to teacher revision

to frequency of practice to frequency of questioning. The combined indirect effects

sum up to a moderate effect size of .21 in mathematics and .23 in English. This again

is an important finding, in that it underscores not only the multiplier effects that

structure and clarity has on other instructional practices, but its ability to help gener-

ate ‘visible’ teaching and learning.

Eighth, a second DI practice, teacher revision, also plays a pivotal role in the inte-

grated model. In mathematics, revision is predicted by three practices: two fraternal

DI practices—MLT (.44) and structure and clarity (.31)—and one first cousin TI

practice, the ubiquitous exam preparation (.23). Teacher revision in turn predicts fre-

quency of questioning through two pathways: directly (.16) and indirectly through

frequency of practice (.17*.34) for a total effect size of .22 for mathematics and .25

for English. In English, we observed some similar but also some unique pathways into

and out of revision. Similar to mathematics, revision is predicted by maximum learn-

ing time (.49) and structure and clarity (.25). However, a unique predictor of revision

found in English, but not mathematics, is memorization (.20). Revision goes on to

predict frequency of questioning directly (.19), but also indirectly through frequency

of practice (.23*.26) for a total effect size of .25. However, we found a significant

effect for a feedback loop connecting revision and textbook focus (.27), and again the

74 D. Hogan et al.

© 2013 British Educational Research Association

Page 19: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

cycle repeats itself—through worksheet focus, drill and memorization, including

some other indirect pathways through structure and clarity and maximum learning

time—resulting in an increased effect size of .30 from revision to frequency of ques-

tioning. This feedback is unique to English and together with the effect of memoriza-

tion on teacher revision which we found earlier, they suggest that perhaps traditional

instruction is more visible as a pedagogical practice among English teachers.

Finally, in mathematics, frequency of practice is predicted by two DI practices and

two TI practices: structure and clarity (.25), teacher revision (.17), textbook focus

(.28) and exam preparation (.14). In turn it has a solitary but relatively substantial

pathway from it to frequency of questioning (.34). In English, it is predicted by three

instructional practices: memorization (.17), textbook focus (.13) and exam prepara-

tion (.15), but it in turn predicts frequency of questioning (.26), indicating that teach-

ers often follow up a practice session with more questioning, surely sensible.

Overall then, the SEM for TI and DI mathematics generates a good fitting model

with a dense, asymmetrical network of pathways leading from the TI ensemble of

instructional practices, exam preparation above all, to the DI instructional practices

with a high value instructional practice, frequency of questioning, as the key outcome

variable. However, the density and strength of these pathways casts considerable

doubt on the idea that we can view TI and DI as discrete instructional categories and

underscores the wisdom of viewing them, at the construct level, as constituting an

integrated, theoretically meaningful hybridic model of instructional practice and that

underscores the exceptionally strong institutional leverage of the assessment system

over instructional practice in Singapore. In English, likewise, the SEM model for TI

and DI is very strong but it also generated a slightly less asymmetrical model between

TI and DI because of the non-recursive feedback from teacher revision to textbook

focus. In addition, English teachers appear to make much greater instrumental use of

memorization focus to leverage additional instructional practices than do mathemat-

ics teachers.

Teaching for understanding (TfU) and co-regulated learning strategies. In this section we

want to consider two quite closely related scales: a teaching for understanding scale

that draws substantially on existing teaching for understanding frameworks, and a co-

regulated learning strategies scale that draws heavily on research on metacognitive

self-regulation. Initially, we considered the latter an aspect of the former, but confir-

matory factor analysis and a number of theoretical considerations convinced us to

consider them separately, although the two are, if not identical twins, closely related

siblings, theoretically speaking.

What we have called the teaching for understanding (TfU) framework shares a

common intellectual sensibility (if not identical constructs) with Harvard Univer-

sity’s Project Zero’s Teaching for Understanding framework (Perkins, 1993) and

Good and Brophy’s expansive account of ‘teaching for understanding’ that draws in

turn on a variety of intellectual resources particularly constructivist (and particularly

social constructivist) models of knowledge, learning and teaching (Brophy, 2002,

2004; Good & Brophy, 2003, Chapter 10). But we have also drawn on other theo-

retical resources as well: the Understanding by Design framework developed by Wig-

gins and McTighe (2005); the ‘authentic pedagogy’ of Newmann and his

Assessment and the logic of instructional practice 75

© 2013 British Educational Research Association

Page 20: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

colleagues at the University of Wisconsin (Newmann et al., 1995); conceptions of

‘thoughtful discourse’, ‘dialogical teaching’ and ‘understanding talk’ variously

developed by Douglas Barnes (1992, 2008), Robin Alexander (2001, 2008), Neil

Mercer and his colleagues (Mercer, 1992; Mercer & Littleton, 2007; Hodgkinson &

Mercer, 2008), Courtney Cazden (Cazden, 1988), Lauren Resnick (Resnick et al.,

2010), Sarah Michaels and colleagues (Michaels et al., 2002, 2004, 2008),

Nystrand and colleagues (Nystrand et al., 1999, 2001) and John Hattie’s recent and

important account of ‘active teaching’ and ‘visible learning’ (2009, 2012). Our

attention to co-regulated learning strategies (CRLS) reflects a research judgment

reached over the past couple of decades that metacognitive self-regulation is a

strong determinant of academic achievement and a key capacity for successful nego-

tiation of twenty-first century institutions (OECD, 1999; Good & Brophy, 2003;

Galton, 2007; James et al., 2007; Hacker et al., 2009; Hattie, 2009, 2012; Zimmer-

man & Schunk, 2011).

Our initial specification of the teaching for understanding (TfU) framework

focused on the following 12 scales, subsequently reduced to 11 when we combined

the 3rd and 4th indicators because of very high collinearity:

(1) An effective focus on developing on understanding (six indicators; e.g., ‘The tea-

cher’s explanations really help me understand the topic’; ‘Class discussions

really help me understand the topic’).

(2) Asking high quality questions designed to prompt students to think about their

learning (four indicators; e.g., ‘The teacher asks good questions to see if we

really understand’; ‘The teacher’s questions help us to think deeply’; ‘The tea-

cher asks lots of questions that open up discussion’).

(3) Communicating learning goals (four indicators; e.g., ‘The teacher tells us the learn-

ing objectives of the lesson’).

(4) Communicating performance standards (four indicators; e.g., ‘The teacher explains

the standard of good performance in our tests and exams’).

(5) Actively attempting to engage students in the work of the class by exciting student

interest and curiosity (five indicators; e.g., ‘The teacher makes mathematics/

English really interesting’).

(6) Flexible teaching: adjusting instructional methods when appropriate as the task or

levels of student engagement require (three indicators; e.g., ‘The teacher tries

different kinds of teaching to help us understand better’; ‘The teacher changes

the speed of the lesson’).

(7) Engaging students in meaningful class discussions (one indicator; e.g., ‘The tea-

cher supports long class discussions about topics’).

(8) Support for collaborative group work (two indicators; e.g., ‘The teacher encour-

ages students to work as a team in group work’).

(9) Teacher scaffolding of group work to ensure that students work together collabora-

tively rather than simply alongside each other (three indicators; e.g., ‘The

teacher shows us how to work together in groups’).

(10) Continuous teacher monitoring of student learning—i.e., feedback from the stu-

dent to the teacher (four indicators; e.g., ‘The teacher asks the class questions to

see how well we understand the topic at the beginning of the class’).

76 D. Hogan et al.

© 2013 British Educational Research Association

Page 21: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

(11) Regular personal feedback from teachers to students about the quality of their

work (five indicators; e.g., ‘The teacher gives me personal comments on my

homework’).

(12) Regular collective feedback from teachers to students about the quality of their

work (five indicators; e.g., ‘The teacher gives the class detailed comments on

exams or tests’).

We report the results of the TfU CFAs for mathematics and English in Table 8. Fit

statistics are very good for both subjects. In addition, in both subjects, quality of ques-

tioning and focus on learning have the highest factor loadings. This is an especially

important finding, indicating that both of these constructs are close to the centre of

the covariance structure of the larger TfU construct. The strength of flexible teaching

is also notable, emphasizing the pivotal importance of flexible teaching to the TfU

framework and supporting the decision to include it in the TfU scale rather than the

DI scale. The strength of the monitoring of student learning aligns very well with the

arguments of Hattie (2009, 2012) and others regarding the critical contribution of

monitoring to visible teaching and learning. Factor correlations fall within an accept-

able range (Table 10), except for two (focus on learning and quality of questioning,

and monitoring student learning and quality of questioning) which are both high and

Table 8. CFA higher order factor loadings and goodness-of-fit statistics: teaching for

understanding, Secondary 3 mathematics and English

Mathematics English

Respecified higher order model Respecified higher order model

Higher-order factor loadings:

Quality of questioning (4) .906 (.011) .919 (.010)

Focus on learning (6) .882 (.013) .933 (.010)

Communicating learning

goals and performance

standards (8)

.873 (.011) .885 (.011)

Monitoring student

learning (4)

.851 (.013) .880 (.014)

Flexible teaching (2) .820 (.019) .854 (.019)

Collective feedback (4) .572 (.024) .487 (.028)

Collaborative group

work (2)

.570 (.027) .699 (.022)

Curiosity and interest (4) .565 (.024) .518 (.026)

Teacher scaffolding of

group work (3)

.549 (.024) .737 (.019)

Whole class discussion (1) .545 (.026) .686 (.022)

Personal feedback (4) .528 (.026) .427 (.029)

Goodness-of-fit statistics:

Chi-square/df/p-value 1653.505/722/.0000 1359.798/722/.0000

CFI/TLI .966/.964 .974/.972

RMSEA (90% C.I.) .033 (.031–.035) .029 (.027–.032)SRMR .034 .028

Assessment and the logic of instructional practice 77

© 2013 British Educational Research Association

Page 22: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

statistically significant, indicting substantial error covariance. However, we thought

all the error covariances theoretically justified and adjusted for them in the final

respecified model.

In English, the strong factor loadings were led by focus on learning (.933), quality

of questioning (.919), communicating goals and standards (.885), monitoring stu-

dent learning (.880) by and flexible learning (.854).The stronger factor loadings for

focus on learning and quality of questioning in English than in mathematics are espe-

cially notable—and not a little surprising. Inter-factor correlations are all reasonable

(see Table 10), as are error covariances. Finally, apropos our decision to include flexi-

ble teaching in the TfU rather than the DI scale, it is worth noting that flexible teach-

ing has the strongest factor loadings in both mathematics and English.

In Table 9 we report the mean scores and standard deviations for the two TfU

scales in both subjects. Overall, the mean scores for TfU were substantially lower

than the mean scores for TI and DI for both subjects: 3.38 in mathematics, and 3.43

in English. However, within subjects mean scores differences between scales are quite

large. In mathematics, for example, the mean scores range from a low 2.79 for teacher

scaffolding of group work to 3.59 for collective feedback. In English, mean scores var-

ied from a low of 3.21 for whole class discussion to a high of 3.58 for collective feed-

back and 3.55 in communicating goals and standards.

From a theoretical perspective, the relative strength of communicating learning

goals and standards, collective feedback, flexible teaching, and monitoring student

learning all indicate the implicit but relatively substantial commitment of Singapore’s

teachers to a ‘visible’ model of teaching and learning (Hattie, 2009). Indeed, we think

this is a good news story, pedagogically speaking, although the weak score for per-

sonal feedback indicates a major gap in the instructional repertoire of Singapore’s

Table 9. Teaching for understanding, Secondary 3 mathematics and English

Mathematics English

dMean (1–5) SD Mean (1–5) SD

N 1166 1027

Teaching for Understanding scale 3.38 .602 3.43 .564 .09

Collective feedback 3.59 .805 3.58 .766 .01

Communicating learning goals

and performance standards

3.57 .771 3.55 .681 .03

Flexible teaching 3.57 .873 3.47 .829 .12

Monitoring student learning 3.46 .801 3.48 .724 .03

Personal feedback 3.43 .829 3.47 .838 .05

Focus on learning

(understanding)

3.36 .710 3.43 .704 .10

Quality of questioning 3.34 .790 3.41 .733 .09

Curiosity and interest 3.25 .898 3.33 .894 .09

Whole class discussion 2.97 1.040 3.21 .929 .24

Collaborative group work 2.87 .962 3.28 .831 .46

Teacher scaffolding of group

work

2.79 1.023 3.28 .849 .52

78 D. Hogan et al.

© 2013 British Educational Research Association

Page 23: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

Table

10.

Latentco

rrelationmatrix:teach

ingforunderstandingformathem

atics

andEnglish

FT

FOL

QTQ

SCF

MSL

PFB

CFB

CNI

CLGPS

CGW

WCD

Flexibleteach

ing(F

T)

.73

.73

.46

.70

.42

.47

.44

.73

.46

.44

Focu

sonlearning(F

OL)

.78

.79

.47

.74

.46

.52

.50

.80

.48

.43

Quality

ofquestioning(Q

TQ)

.78

.87

.53

.79

.47

.51

.52

.77

.57

.55

Teach

erscaffoldingofgroupwork

(SCF)

.64

.72

.66

.46

.32

.30

.34

.45

.80

.58

Monitoringstuden

tlearning(M

SL)

.78

.79

.81

.65

.44

.47

.46

.74

.48

.46

Personalfeed

back

(PFB)

.41

.37

.38

.33

.39

.81

.72

.47

.32

.31

Collectivefeed

back

(CFB)

.47

.41

.45

.38

.44

.78

.77

.53

.26

.23

Curiosity

andinterest(C

NI)

.49

.47

.47

.36

.46

.73

.78

.50

.34

.33

Communicatinglearninggoalsand

perform

ance

standards(C

LGPS)

.74

.84

.81

.63

.81

.38

.45

.45

.47

.47

Collaborativegroupwork

(CGW

).63

.66

.64

.82

.62

.35

.35

.38

.60

.77

Wholeclass

discu

ssion(W

CD)

.59

.66

.65

.62

.58

.33

.34

.38

.57

.69

Note:Allco

rrelationssignificantatp<.01.Correlationsformathem

atics

are

totheleftofthediagonalandEnglish

totheright.

Assessment and the logic of instructional practice 79

© 2013 British Educational Research Association

Page 24: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

teachers in both English and mathematics. More broadly, the relatively low scores for

focus on learning and quality of questioning are especially troubling while the particu-

larly low scores for collaborative group work, whole class discussion, scaffolding of

group work, curiosity and interest and personal feedback points to critical weak spots

from a TfU perspective, particularly with regard to the value of classroom talk as a

means of co-constructing meaning and developing conceptual understanding.

Correlations between the latent TfU measures are reported in Table 10. In gen-

eral, in both subjects, these are moderate to large. The strength of the correlations

between focus on learning and the remaining TfU variables, is particularly notable

and affirmative of the empirical, as well as theoretical, warrant for the TfU scale. We

consider the SEMmodels for TfU and CRLS jointly in light of their intimate concep-

tual (and, as it turns out, empirical) relationship (Figures 7 and 8).

As we have seen, our TfU scale is based on 11 separate (but relatively highly inter-

correlated) subscales. The co-regulated learning strategies scale (CRLS) scale is a

second order construct that builds on three multi-item first order scales for self-direc-

ted learning, self-assessment, and peer assessment and draws on recent theoretical lit-

erature (National Research Council, 2000, pp. 14–18; Paris & Paris, 2001; Nota et

al., 2004; Mok et al., 2006; James et al., 2007, pp. 3, 5; Darling-Hammond, 2008;

Hacker et al., 2009; Hattie, 2009; Zimmerman & Schunk, 2011). Meanwhile, a num-

ber of researchers have given considerable thought to how students might learn meta-

cognitive self-regulation. Dylan Wiliam, for example, argues that peer and self-

assessment is a particularly effective mechanism for developing the classroom as a co-

regulated learning environment (CRLE) and cultivating metacognitive self-regulation

and student management of their learning (Black et al., 2003, pp. 49–56; Wiliam,

2007, pp. 1076–1084).As an example, our specification of the CRLS and the CFA for the three compo-

nent sub-scales for English are reported below in Table 11. The models generated

Note: All values represent unstandardized estimates significant at p < .01. In brackets are standard errors. Double-headed curved arrows denote correlated latent disturbances.

Goodness-of-fit statistics: mathematics(N = 1166)Chi-square/df/p-value 75.248/49/.0094CFI/TLI .997/.996RMSEA (90% CI) .021 (.011–.03)SRMR .015

.14(.03)

.70(.04)

.43(.04)

.20(.04)

.23(.05)

.35(.05)

.36(.05)

.28(.06)

.28(.04) .43(.04)

.49(.06)

.37(.04)

.40(.04)

Monitor Student Learn

Collec ve Feedback

Focus on Learning

Comm Goals/stan

Teacher Scaffolding

Quality of Ques on

Flexible Teaching

Curiosity & Interest

Personal Feedback

.31(.04)

.24(.05)

.63(.04)

.57(.04)

.11(.03)

Whole Class Disc

Collabora ve GpWk

.55(.03)

.61(.04)

.12(.03)

.12(.03)

.25(.04)

CRLS .63(.04)

.19(.03)

.15(.04)

.13(.03)

Figure 7. SEMmodel for Teaching for Understanding and co-regulated learning strategies:

mathematics

80 D. Hogan et al.

© 2013 British Educational Research Association

Page 25: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

strong goodness-of-fit statistics and factor loadings. All three scales are also highly—but not too highly—correlated. Factor loadings are all quite high, with one especially

high (personal assessment in English), a result that is not statistically troublesome but

that suggests that at least in English, personal assessment is an exceptionally proxy

for, and is indistinguishable from, the higher order latent construct (CRLS).

Table 12 reports the mean values, standard deviations and effect sizes of our mea-

sures for CRLS. What they indicate is that in 2010, teachers provided moderate

opportunities for self-directed learning, but far fewer opportunities for self-assess-

ment or peer assessment in both mathematics and English. The findings also indicate

that opportunities for CRLS are stronger in Secondary 3 English classes than in Sec-

ondary 3 mathematics classes.

Figures 7 and 8 report the TfU/CRLS SEM models for mathematics and English

respectively. (We also ran the TfU models separately with focus on learning rather

than CRLS as the key outcome variable, but since the results are almost identical to

the combined TfU/CRLS models, we have decided to report on the latter only here.)

Both models have exceptional good fit, rich, sensible and suggestive networks of path-

ways, strong coefficients, and theoretical gravitas. The selection of focus on learning

as the key outcome variable in the TfU models reflects our argument earlier that it

indexes the degree to which teachers focus on their instruction on meaning making

and developing student understanding at the heart of the TfUmodel generally. CRLS

as an outcome measure, on the other hand, identifies the extent to which teachers

seek to give students the opportunity to manage their own learning in line with the

substantial research on metacognition that underscores its importance in student

learning. In both subjects CRLS is positioned as an endogenous outcome measure

rather than a mediating variable that predicts variation in focus on learning. We also

attempted to see whether we could position CRLS as a predictor of focus on learning,

Note: All values represent unstandardized estimates significant at p < .01, *p < .05. Double-headed curved arrows denote correlated latent disturbances.

Goodness-of-fit statistics: English(N = 1027)Chi-square/df/p-value 74.146/52/.0235CFI/TLI .997/.996RMSEA (90% CI) .020 (.008–.030)SRMR 0.17

.26(.04)

.65(.04)

.52(.06)

.28(.06)

.16(.05).38(.05)

.32(.05)

.30(.05)

.40(.05) .26(.05)

.53(.05)

.29(.06)

.28(.06)

Monitor Student L

Collec ve Feedback

Focuson Learning

Comm Goals/stan

Teacher Scaffolding

Quality of Ques on

Flexible Teaching

Curiosity & Interest

Personal Feedback

.20(.05)

.40(.04)

.80(.04)

.65(.04)

Whole Class Disc

Collabora ve GpWk

.47(.07)

.43(.04)

.16(.03)

.43(.04)

.25(.07)

CRLS

.87(.05)

*.10(.05)

Figure 8. SEMmodel for Teaching for Understanding and co-regulated learning: English

Assessment and the logic of instructional practice 81

© 2013 British Educational Research Association

Page 26: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

Table 11. CFA co-regulated learning in Secondary 3 English

Variable

Standardized parameter estimates

Initial model Respecified model

11 items 10 items

Self-directed learning

The teacher encourages us to set

our own learning goals

.755 (.013) .758 (.013)

The teacher encourages us to

identify strategies to achieve our

learning goals

.823 (.011) .821 (.011)

The teacher encourages us to

check frequently that our work is

acceptable

.764 (.013) .764 (.013)

Personal assessment

The teacher asks us to grade our

own work

.729 (.013) .733 (.013)

The teacher explains how we can

grade our own work

.788 (.011) —

The teacher expects us to discuss

our own grading of our own work

.796 (.011) .813 (.011)

The teacher encourages us to

comment on our own work

.798 (.011) .810 (.011)

Peer assessment

The teacher asks students to grade

each other’s work

.815 (.010) .816 (.010)

The teacher explains how we can

grade each other’s work

.805 (.010) .800 (.010)

The teacher expects us to discuss

our grading of each other’s work

.825 (.010) .827 (.010)

The teacher encourages us to

comment on each other’s work

.810 (.010) .812 (.010)

Latent correlations:

Self-directed learning↔ personal

assessment

.851 (.012) .818 (.014)

Self-directed learning↔ peer

assessment

.666 (.018) .666 (.018)

Personal assessment↔ peer

assessment

.799 (.013) .783 (.014)

Higher order factor loading:

Self-directed learning .842 (.013) .834 (.014)

Personal assessment 1.01 (.010) .981 (.012)

Peer assessment .791 (.014) .798 (.014)

Goodness-of-fit statistics:

Chi-square/df/p-value 230.847/41/.000 138.695/32/.000

CFI/TLI .983/.977 .989/.985

RMSEA (90% C.I.) .052 (.045–.058) .044 (.037–.052)SRMR .021 .018

Note: The factor loadings were fixed to equality as the standardized estimates for personal assessment in initial

model was > 1.0.

82 D. Hogan et al.

© 2013 British Educational Research Association

Page 27: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

on the grounds that teachers might use CRLS to enhance their focus on learning. But

the goodness-of-fit statistics with CRLS as a predictor variable is not as strong as the

excellent goodness-of-fit statistics with CRLS as the outcome variable in the two

models (in mathematics, a chi-square of 180.349/49/.000 versus 75.248/49/.094; in

English, 178.412/52/.000 versus 74.146/.52/.0235).

Before deciding on the final models, we tried a range of alternative models with dif-

ferent exogenous constructs, including flexible teaching and monitoring student

learning, but could not make the models work as well (as measured by the goodness-

of-fit statistics). Beyond this, a number of features of the two models stand out.

We begin with the two key outcome measures: co-regulated learning strategies

(CRLS) and focus on learning (FOL). But before we do, it’s important to recall that

in mathematics, the mean score for CRLS is a very low 3.01, in English, 3.28, indicat-

ing that teachers do not employ CRLS particularly often, especially in mathematics.

Nor does the use of CRLS, such as it is, alter the pattern and strength of the overall

structure of TfU practices in any obvious way in either of the two subjects if we com-

pare the two TfU SEMS with the CRLS SEMS. Instead, the internal structure of the

two models remains remarkably stable. In mathematics, CRLS is predicted multivari-

ately by teacher scaffolding of group work (.19), by whole class discussion (.13), by

focus on learning (.63) and by communication of learning goals and performance

standards (.15). All these are theoretically sensible. Of these, by far the strongest and

most interesting is focus on learning, for what this pathway suggests is that teachers

do not so much employ CRLS to increase their focus on learning but use a focus on

learning to scaffold a transition to CRLS. The one construct missing that we also

expected to predict CRLS is collaborative group work, although there is an indirect

pathway from it through teacher scaffolding of group work (.70*.13 = .09). In Eng-

lish, there are only two statistically significant pathways into CRLS: focus on learning

(.87) and flexible teaching (.10). Both of these are theoretically sensible. Indeed, all

told, the source of the strongest indirect effect on CRLS in both subjects is flexible

teaching.

We positioned focus on learning as the key endogenous outcome measure in the

two TfU models and as the penultimate outcome measure in the two CRLS models.

In both the TfU and CRLS models focus on learning (FOL) shares three predictors

in both subjects: communication of learning goals and standards (.36 and .32 respec-

Table 12. Co-regulated learning strategies, Secondary 3 English and mathematics

Mathematics English

dMean (1–5) SD Mean (1–5) SD

Co-regulated learning

strategies (Alpha =.918, .920)

3.01 .770 3.28 .688 .37

Self-directed learning 3.41 .794 3.45 .747 .05

Self-assessment 2.92 .907 3.20 .782 .33

Peer assessment 2.80 .945 3.23 .802 .49

Assessment and the logic of instructional practice 83

© 2013 British Educational Research Association

Page 28: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

tively), quality of questioning (.35 and .38), and flexible teaching (.23 and .16). In

addition, in mathematics, for example, there are indirect pathways from flexible

teaching to focus on learning via quality of questioning (.20*.35 = .07) and commu-

nicating learning goals (.37*.36 = .133), while in English there are, for example, the

following indirect pathways: one via quality of questioning, whole class discussion,

collaborative group work and teacher scaffolding (.28*.47*.43*.65*.16 = .005), one

through whole class discussion, collaborative group work and teacher scaffolding

(.25*.43*.65*.16 = .011), and the third through collaborative group work, teacher

scaffolding of group work and focus on learning (.43*.65*.16 = .04). While the indi-

rect effects are small, they are theoretically sensible.

Flexible teaching (FT) functions as a critical lynchpin in all four TfU/CRLS mod-

els. We know from Table 9 that the mean scores for FT for both mathematics and

English are moderately high (3.57, 3.47 respectively), relatively speaking, although

against informal statistical protocols, these means are not especially high. Still,

although it has only one pathway leading into it (curiosity and interest with a respect-

able coefficient of .24), in mathematics, it has four pathways leading out of it: moni-

toring student learning (.63), communication of learning goals and standards (.37),

focus on learning (.23), quality of questioning (.20). In English, there are seven path-

ways emanating from FT: monitoring student learning (.80), communication of

learning goals and standards (.29), focus on learning (.16), quality of questioning

(.28), whole class discussion (.25), collaborative group work (.43) and self-directed

learning (.10). Not all of these are substantial coefficients, but one is very substantial

and the others moderate to strong. We interpret the striking generativity of FT, espe-

cially in English, to signify the pragmatic, non-sectarian instrumentalism of Singapo-

rean instruction.

For some researchers, feedback promises to be the Holy Grail of instructional prac-

tice. This might well prove to be the case but it’s clear from the mean scores reported

in Table 9 that teachers in both mathematics and English provide relatively substan-

tial levels of collective feedback to their students (3.59 and 3.58 for collective feed-

back respectively), but lower levels of personal feedback to their students (3.43 and

3.47 respectively). This suggests room for improvement, particularly in the provision

of personal feedback, given its demonstrable ability to enhance student understanding

and learning (Black & Wiliam, 1998; Hattie, 2009). This judgement is reinforced by

the fact that both forms of feedback generate only one pathway each—and both to

curiosity and interest (.49 and .28 respectively in mathematics, and .53 and .30

respectively in English). These are significant coefficients, nearly identical across sub-

jects, and theoretically sensible, since we can easily understand why teachers would

deliberately use feedback to students to build interest and curiosity among them. The

oddity is that the coefficients are stronger for collective feedback rather than personal

feedback, suggesting, given international research findings, that Singaporean teachers

do not fully exploit the ability of personal feedback to build student engagement as

much as they might. There are of course multiple indirect effects from the two forms

of feedback to focus on learning, both going through curiosity and interest, but once

these are multiplied out, the overall effect size is relatively small (.13 and .17 for

mathematics and English for collective feedback, and .08 and .10, respectively, for

personal feedback; Tables 13 and 14). This too reinforces the conclusion that the

84 D. Hogan et al.

© 2013 British Educational Research Association

Page 29: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

overall generativity of the two modes of feedback is very low. This in itself does not

challenge the argument of Black and Wiliam and others who have argued for the

unique strength of the impact feedback on student learning given that we are measur-

ing instructional practices rather than student learning, but it reinforces the impres-

sion that Singaporean teachers do not employ feedback as often or as generatively as

they might.

A further feature of feedback practices in both mathematics and English is that

both forms of feedback are predicted by the communication of learning goals and

standards (and with reasonably robust coefficients, especially in mathematics), and in

so-doing enhance the visibility of learning processes in the classroom. However, there

are no pathways from the monitoring of student learning to either of the feedback

constructs in both models. This is, to put it mildly, paradoxical, for we would nor-

mally expect teachers to provide feedback only after they have assessed the state of

student learning in their classroom, not without it. What the absence of these path-

ways suggests is that teachers do not follow up monitoring with feedback practices.

However, given the nature of the data we cannot draw firm conclusions on this mat-

ter—it merely indicates a suboptimal pattern of instructional behavior that we will test

further with qualitative data in a later publication. We will also use qualitative data to

follow up on the intriguing pathway from collaborative group work to personal feed-

back in mathematics. Although small (.12), it is statistically significant and intriguing,

suggesting teachers occasionally follow up group work with personal feedback to indi-

vidual students within the group.

We expected that curiosity and interest would exhibit a dense network of pathways

backwards and forwards, but we did not find this to be the case. In both mathematics

and English, personal and collective feedback predicted curiosity and interest,

although the coefficients for collective feedback are far higher (.49 and .53 respec-

tively) than they are for personal feedback (.28, .30 respectively). It is also the case

that in both subjects, teachers provide collective feedback more often than they do

personal feedback. In effect, teachers are far more likely to follow up collective feed-

back with activities that engage student interest and curiosity than they do following

personal feedback. Normatively, this is not optimal, given that research findings gen-

erally suggests that personal feedback, appropriately given, is far more likely to

enhance student learning than collective feedback (Hattie, 2009). At the same time,

however, the generativity of curiosity and interest is quite limited in both subjects. In

mathematics, there are two pathways from curiosity and interest, one to flexible

teaching (.24), one to quality of questioning (.11). Apparently, teachers respond to

student curiosity and interest by becoming more flexible and asking more demanding

questions. We expected to find that the pathways would go in the opposite direction,

but when tested, they were not statistically significant. In English there is only one

generative pathway—to flexible teaching (.40).

Whereas we included a measure of the frequency of questioning (FOQ) in the

Direct Instruction scale, we included a measure for quality of questioning (QTQ) in

the TfU scale designed to capture whether or not the questions teachers ask their stu-

dents develop their understanding of the topic or task (as in, for example, ‘the teacher

asks good questions to see if we really understand’). The news on the QTQ front is

mixed. The bad news is that the mean scores for the QTQ scale in both subjects are

Assessment and the logic of instructional practice 85

© 2013 British Educational Research Association

Page 30: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

relatively low (3.34 and 3.41 respectively; Table 9). The good news is that QTQ is

heavily enmeshed in a network of instructional practices. On the one hand, it is pre-

dicted by communication of learning goals and standards in both mathematics and

English (.28 and .40 respectively), flexible teaching (.20 and .28) and student moni-

toring (.28 and .40). Curiosity and interest also predicts QTQ but in mathematics

only (.11). On the other hand, QTQ predicts other instructional practices in turn:

focus on learning (.35 and .38), whole class discussion (.55 and .47), collaborative

group work (mathematics only, .25) and teacher scaffolding of group work (.14, .26).

In effect, the quality of teacher questioning matters, and matters a lot, especially

where it counts in generating whole class discussion and a focus on learning.

We suggested earlier, following John Hattie’s work, that communication of learning

goals and performance standards (CLGPS) is a key indicator of ‘visible learning’ that

he found to be highly predictive of successful teaching and learning. We also indi-

cated that relative to other instructional practices, Singaporean teachers do reasonably

well in this regard, with a mean score of 3.57 in mathematics and 3.55 in English.1.

CLGPS is predicted by monitoring of student learning (.43 and .52 respectively in

both mathematics and English), and by flexible teaching (.37 and .29 respectively).

These coefficients are both substantial and theoretically sensible: the first indicates

that teachers often respond to their monitoring of student learning with statements or

reiterations of their learning goals and performance standards, while the latter sug-

gests that teachers, as a part of their repertoire of effective teaching, state or reiterate

their learning goals and performance standards for the lesson. However, the SEM

model indicates that teacher communication of learning goals and performance stan-

dards is not especially generative in its own right, given that it generates only two

direct pathways in each subject: to focus on learning (.36 and .32 respectively) and to

quality of questioning (.28 and .40 respectively). However, there is an indirect path-

way from communication of learning goals and performance standards to focus on

learning via quality of questioning in both mathematics (.28*.35 = .10) and English

(.40*.38 = .152), along with, for example, indirect pathways in English via quality of

questioning, whole class discussion, collaborative group work and and teacher scaf-

folding of group work.

There is by now a substantial body of academic research that has focused on the

educational value of appropriately structured classroom dialogue—see, for example,

Barnes (1992, 2008), Alexander (2001, 2008, 2012), Mercer and his colleagues

(Mercer, 1992; Mercer & Littleton, 2007; Hodgkinson & Mercer, 2008), Cazden

(1988), Resnick (Resnick et al., 2010), Nystrand and colleagues (Nystrand et al.,

1999, 2001), Michaels and colleagues (Michaels et al., 2002, 2004, 2008) and Hogan

et al. (2012b). While our measure of whole class discussion does not specify key indi-

cators of ‘dialogue’ per se it does indicate in a very general way the frequency with

which teachers allow or support extended class discussions of various topics. The

mean scores reported in Table 9 indicate, however, that teachers do not permit

extended whole class discussions very often in either subject. In mathematics, the

mean score is a very low 2.97, and in English, a slightly better 3.21, but relative to

other TfU instructional practices in English, it is the instructional practice with the

lowest mean score in the larger TfU inventory. The SEM models also indicate that

whole class discussion is only weakly embedded in the network of TfU/CRLS

86 D. Hogan et al.

© 2013 British Educational Research Association

Page 31: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

instructional practices in the two subjects. In mathematics, whole class discussion is

predicted by the quality of teacher questioning (.55) and nothing else. In English, it is

predicted by the quality of teacher questioning (.47) and by flexible teaching (.25),

suggesting that whole class discussion is a standard element of the flexible teaching

repertoire, at least in English, as well as a sensible follow-up activity that teachers

employed after asking high quality questions. Yet, surprisingly, whole class discussion

is not particularly generative in its own right in either subject, leading only to collabo-

rative group work (.61, .43 respectively, in mathematics and English).

The last two—and closely related—constructs we want to discuss focus on collabo-

rative group work and teacher scaffolding of group work. The former construct mea-

sures the frequency with which teachers ‘breaks the students up into small groups to

work together’. The latter focuses on the efforts teachers make to structure group

work in a way that builds an effective division of labor with a high level of interdepen-

dency by showing ‘us how to work together in groups’. While asking students to work

in groups collaboratively is all well and good, as Galton (2007) and others have

reminded us over many years, without careful teacher scaffolding of group work the

students are likely to end up working individually in pseudo groups rather than collec-

tively and collaboratively. But the theoretical promise is high; the mean scores for

both sets of constructs in both subjects are very low, especially in mathematics (2.87

and 2.79). In English, the mean scores are a little higher for both constructs (3.28

and 3.28). In short, mathematics and English teachers in Singapore, especially math-

ematics teachers, make little use of collaborative group work in their classes. In addi-

tion, there are only very limited pathways into and out of the two constructs. In

mathematics, collaborative group work is predicted by whole class discussion (.61)

and by the quality of questioning (.25). In English, whole class discussion also pre-

dicts collaborative group work (.43), but not quality of questioning. Instead, collabo-

rative group work is also predicted by that highly energetic TfU pivotal construct,

flexible teaching (.43). In effect, in both mathematics and English, teachers often

segue from whole class discussion into small group work, but in mathematics, teach-

ers often follow up demanding questions with collaborative group work, whereas in

English, collaborative group work appears to be a more basic staple of teachers’

instructional repertoire. In both mathematics and English, teacher scaffolding of

group work is predicted, reasonably enough, by collaborative group work itself (.70

and .65 respectively) and, perhaps less predictably, occasionally by the quality of

questioning (.14 and .26), suggesting that teachers occasionally—very occasionally—use high quality questions to set up the structure of the group work. There is also a

weak pathway from collaborative group work to personal feedback (.12) in mathe-

matics, indicating that teachers sometimes provide personal feedback on the basis of

their work in small groups, at least in mathematics. The generativity of teacher scaf-

folding though is very limited, with no pathways leading from it to other instructional

practices in mathematics and only one weak pathway, from teacher scaffolding to

focus on learning (.16), in English.

Tables 13 and 14 report the direct, indirect and total effects of the pathways

mapped in Figures 7 and 8 assuming focus on learning (FOL) as the endogenous out-

come variable. For both mathematics and English, flexible teaching has by far and

away the strongest total effects (.74 and .84). Communication of learning goals and

Assessment and the logic of instructional practice 87

© 2013 British Educational Research Association

Page 32: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

performance standards (.54) provides the second strongest effect size in mathematics

and English, while the monitoring student learning (MSL) (.37) and the quality of

questioning (QTQ) (.46) is the third strongest effect in mathematics and English

respectively. In broad terms then, the generativity of flexible teaching, followed by the

communication of learning goals and performance standards, in the Singapore con-

text is consistent with the theoretical arguments we endorsed earlier regarding the

importance of instructional alignment and visible learning. These findings do not of

course demonstrate that these instructional practices optimize student learning; they

are merely consistent with the broad theoretical claims made on their behalf by

researchers. We will report their impact on student learning in a latter report. For

now though, our modelling of TFU instructional practices indicates that instructional

practices in both mathematics and English are broadly in line with the normative

injunctions that researchers have identified, but that there are a number of practices

that are clearly suboptimal, normatively speaking, including, in particular, the

absence of pathways from monitoring student learning and feedback, and from whole

class discussion and collaborative group work to focus on learning.

Table 15, similarly, reports the direct, indirect and total effect sizes for the two

SEM models with co-regulated learning strategies (CRLS) as the key outcome vari-

able. Again, flexible teacher leads the way, followed by monitoring student learning.

Hybridity, assessment and the logic of instruction

In Table 16 we summarize the mean scores for the four instructional strategies

reported above. Clearly, the rank order differs in the two subjects and the spread is

unequal. In mathematics, TI leads over DI, followed by TfU a fair way back, and

CRLS in that order. The spread is quite substantial between TI/DI and TfU, and

even bigger between TfU and CRLS. In English, DI leads over TI, followed by TfU

and CRL in that order. The spread between the scales is not as wide in English as it is

in mathematics. Effect sizes are generally quite large. Moreover, at the scale level,

there are clear differences in mean scores between the two subjects, particularly with

respect to TI and CRLS, and these are reflected in the effect sizes. Although

the strength of TI in mathematics might lead one to conclude that mathematics

Table 13. Total indirect and direct effect sizes, SEM TfUmodel, mathematics

Pathway

Direct, indirect and total effects

Indirect Direct Total

FT to FOL .50(.04)* .24(.05) .74(.04)*

CLGPS to FOL .17(.02)* .37(.05) .54(.05)*

MSL to FOL .37(.04)* — .37(.04)*

QTQ to FOL — .34(.05) .34(.05)*

CNI to FOL .26(.04)* — .26(.04)*

CFB to FOL .13(.02)* — .13(.02)*

PFB to FOL .08(.02)* — .08(.02)*

Notes: *Significant at p < .01;— refers to paths not modelled.

88 D. Hogan et al.

© 2013 British Educational Research Association

Page 33: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

instruction at least conforms to the East Asian stereotype, the relative strength of the

other instructional strategies should give pay to that suggestion.

A second look at Table 16 reveals a relatively narrow spread between the mean

scores in each subject, apart from CRLS. This hints at the possibility that the leading

three instructional strategies covary together. And indeed, this conclusion is sup-

ported by the correlations reported in Table 17: the correlations between DI, TI and

TfU are all high, whereas the correlations of each of these with CRLS are substan-

tially lower. This is an important finding, for it indicates some differentiation between

strategies that give the teacher an active instructional role in the classroom—TI, DI

and TfU—and the one strategy in our taxonomy (CRLS) that clearly shifts the centre

of instructional gravity away from teachers towards students and student agency. Just

as importantly, the strength of the correlations suggests that Singaporean teachers in

both English and mathematics appear to draw upon all three teacher-focused instruc-

tional strategies jointly rather than choose between them. We doubt very much that

they do so indiscriminately. Rather, our qualitative evidence, based on interviews

with 115 teachers, suggests that they do so because of their strikingly instrumentalist

Table 14. Total indirect and direct effect sizes, SEMTfUmodel, English

Pathway

Direct, indirect and total effects

Indirect Direct Total

FT to FOL .67(.05)* .17(.05)* .84(.04)*

CLGPS to FOL .25(.03)* .29(.05)* .54(.05)*

QTQ to FOL .06(.02)* .40(.06) .46(.06)*

MSL to FOL .42(.05)* — .42(.05)*

CNI to FOL .33(.04)* — .33(.04)*

CFB to FOL .17(.03)* — .17(.03)*

SCF to FOL — .16(.03)* .16(.03)*

PFB to FOL .10(.02)* — .10(.02)*

WCD to FOL .04(.01)* — .04(.01)*

CGW to FOL .11(.02)* — .11(.02)*

Notes: *Significant at p < .01;— refers to paths not modelled.

Table 15. Total indirect effect sizes, SEMTfU-CRLS model, mathematics and English

Pathway

Direct, indirect and total effects

Mathematics English

Indirect Direct Total Indirect Direct Total

FT to CRLS .70(.03)* — .70(.03)* .74(.05)* .10 (.05)^ .84(.04)

MSL to CRLS .41(.03)* — .41(.03)* .37(.05)* — .37(.05)*

CNI to CRLS .26(.04)* — .26(.04)* .33(.04)* — .33(.04)*

CFB to CRLS .13(.02)* — .13(.02)* .18(.03)* — .18(.03)*

PFB to CRLS .07(.02)* — .07(.02)* .10(.02)* — .10(.02)*

Notes: *Significant at p < .01; ^Significant at p < .05;— refers to paths not modelled.

Assessment and the logic of instructional practice 89

© 2013 British Educational Research Association

Page 34: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

and performative orientation to teaching and learning, prompting them to adopt a

pragmatic, fit-for-purpose hybridic instructional practice focused on exam perfor-

mance. At a system level, this suggests a well-integrated, coherent pedagogy that is

not at war with itself or riven by deep divisions on matters of pedagogical faith and

doctrine. Given the extraordinary success of Singaporean schools in international

assessments, such pedagogical hybridity appears to work.

Nonetheless, while the commitment of Singaporean teachers to traditional and

direct instruction, together with a milder commitment to elements of a teaching for

understanding instructional strategy, might suggest that teachers in Singapore lean

towards, to use John Hattie’s nomenclature, a model of ‘active teaching’, this would

be mildly misleading. This is not because teachers favour ‘constructivist’ or ‘facilita-

tive teaching’. They most definitely do not. While teachers are clearly active in Sin-

gaporean classrooms, they are not particularly active in the ways that matter, or at

least in the ways that John Hattie thinks matter. For Hattie, active teaching supports

‘visible learning’ and ‘is much more effective than unguided, facilitative instruction’

in promoting student achievement. ‘It is essential’, he writes:

… to have visible teaching and visible learning. This notion encapsulates directive, activat-

ing, and involved sets of actions and content, working with students so that their learning

is visible such that it can be monitored, feedback provided, and information given when

learning is successful. (Hattie, 2009, p. 37)

Elsewhere, we have argued that Hattie’s insistence on a combination of active

teaching and visible learning represents a sensible, evidence-based instructional

framework, but that his distinction between active teaching and constructivist teach-

ing is overdrawn and that his account of visible learning and teaching could be

enhanced by taking into account the importance of epistemic clarity—clarity about

the nature of the knowledge work that students engage in—in the design and imple-

mentation of instructional tasks during lessons (Putnam et al., 1990; Schoenfeld,

1992; Stein et al., 1996, 2009; Perkins, 1998; Rittle-Johnson & Alibali, 1999; Schraw,

2006; Hogan et al., 2011, 2012c, e; Rahim et al., 2012).

Despite our reservations about Hattie’s account of active teaching and visible learn-

ing, both conceptions contribute significantly to the scholarship on teaching and

learning. Unfortunately, we do not think the evidence from our broader research pro-

gram indicates that Singaporean teachers are committed to active teaching in ways

that matter. In particular, they rarely employ the kind of instructional strategies that

Hattie positions as central to the active teaching/visible learning nexus. Table 18, for

instance, reports our overall summative judgments from the three projects that make

up our broader Core 2 research program. (Panel 3, for example, is a classroom obser-

vation study of 624 lessons that we also conducted in 2010 using a subsample of clas-

ses and schools from the Panel 2 survey sample reported in this article, while Panel 5

uses the same sample to examine the intellectual quality of tasks and student work.)

It suggests a pretty bleak picture (Hogan et al., 2012d). In addition, evidence from

Panels 2, 3 and 5 reported in Table 19 and represented in Figure 9 indicate that

teachers in Secondary 3 mathematics and English rarely encourage strong forms of

student agency in the classroom, either in their support for self-directed learning, or,

even more fundamentally, in their design of instructional tasks that create substantial

90 D. Hogan et al.

© 2013 British Educational Research Association

Page 35: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

opportunities for students to exercise epistemic, cognitive, metacognitive and discur-

sive agency that are pivotal to knowledge building pedagogies (Hogan et al., 2012d).

In sum, our evidence suggests an instructional framework that is both hybridic in

nature and performative in orientation and with very little commitment to active

teaching and visible learning. These findings again reinforce our conclusion that what

explains the comparative success of Singapore in international assessments is not the

commitment of its teachers to high leverage knowledge practices, active teaching or a

knowledge building pedagogy more broadly, but their single minded commitment

(for many complex reasons) to a highly instrumentalist and performative pedagogy.

Indeed, we can demonstrate this statistically with a SEM model that incorporates all

four instructional strategies—traditional, direct, TfU and CRLS—in a single model.

The SEM models represented in Figures 10 and 11 (and the relevant goodness-of

fit-statistics reported in Table 20) model the overall structure of all four instructional

strategies—traditional, direct, TfU and CRLS—jointly at the construct level in an

integrated model. Overall, keeping in mind that the models are very large and

Table 16. Summary of mean scores, by scale, Secondary 3 mathematics and English

Secondary 3

mathematics Secondary 3 English Effect size

Mean (1–5) SD Mean (1–5) SD d

TI 3.69 .642 3.45 .669 .37

DI 3.67 .670 3.61 .655 .09

TfU 3.38 .602 3.43 .564 .09

CRLS 3.01 .770 3.28 .688 .37

d d

TI vs DI .03 .24

TI vs TFU .50 .03

TI vs CRLS .96 .25

DI vs TfU .46 .30

DI vs CRLS .91 .49

TfU vs CRLS .54 .24

Table 17. Correlation matrix: instructional methods Secondary 3 mathematics and English

TI DI TfU CRLS

Mathematics

Traditional instruction 1

Direct instruction .72** 1

Teaching for understanding .58** .70** 1

Co-regulated learning strategies .28** .35** .73** 1

English

Traditional instruction 1

Direct instruction .75** 1

Teaching for understanding .63** .68** 1

Co-regulated learning strategies .41** .39** .77** 1

Assessment and the logic of instructional practice 91

© 2013 British Educational Research Association

Page 36: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

complex, the goodness-of-fit statistics are exceptionally good. Notably, the chi-square

statistic for English is substantially lower than it is for mathematics (353.981/194/

.0000 versus 404.786/191/.000). Both models are fully recursive: above all, there are

no feedback loops from TfU back to TI or DI practices. In addition, the internal

structure of each of the broader instructional categories (as indicated by the pattern

of pathways between practices within each category) remained remarkably stable.

Task SetUp and Enactment, Student Agency and Knowledge Building

Epistemic Agency

Discursive Agency

Cognitive Agency

Metacognitive Agency

Task Set Up andEnactment: DistributingOpportunities to Learn

(Opportunities toGenerate,

Communicate,Deliberate and Justify

Knowledge Claims)

(Opportunities todevelop Conceptual

Understanding)

(Opportunities toacquire Metacognitive

Knowledge andExercise MC Self-

Regulation)

(Opportunitiesfor Extended

UnderstandingTalk)

Figure 9. Task set-up and enactment, student agency and knowledge building

Note: All values represent unstandardized es mates significant at p<.01. In brackets are standard errors. Double-headed curved arrows denote correlatedlatent disturbances.

Note. EXPR=Exam prepara on; TXBF=Textbook Focus; WKSF=Worksheet Focus; DRILL=Drill; MEM=Memoriza on; SNC=Structure & Clarity; MLT=Maximumlearning me; REV=Revision; FOP=Frequency of prac ce; FOQ=Frequency of ques oning; CFB=Collec ve feedback; PFB=Personal feedback; CNI=Curiosity &

interests; FLT=Flexible teaching; MSL=Monitoring student learning; CLGPS=Communica ng learning goals and performance standards; QTQ=Quality ofques oning; WCD=Whole class discussion; CGW=Collabora ve group work; SCF=Teacher scaffolding; FOL = Focus on learning; CRLS=Co-regulated Learning

MSL

CFB

FOLCLGPS

SCFQTQFLT

CNI

PFB

WCD CGW

CRLS

EXPR REV

SNC

MLT

FOQ

FOPTXBF

DRILL

WKSF

MEM

.33(.06)

.19(.04).34(.04)

.24(.05).14(.04)

.46(.06)

.48(.04)

.17(.04).46(.05)

.17(.05).33(.05)

.13(.05)

.17(.05)

.19(.06)

.25(.06).27(.05).14(.05)

.57(.04)

.43(.04)

.33(.05)

.22(.04)

.27(.04)

.29(.04)

.47(.04)

.12(.04)

.31(.06)

.15(.05)

.43(.05)

.17(.05)

.37(.06)

.59(.04).14(.04)

.15(.04) .39(.04)

.34(.04)

.22(.03)

.20(.04)

.28(.04).40(.04)

.12(.03)

.24(.05)

.36(.05)

.35(.05)

.14(.03)

.70(.04)

.61(.04)

.25(.04).56(.03)

.64(.04)

.19(.03)

.15(.04)

.13(.03)

.27(.05)

.13(.03)

.33(.05)

.25(.04)

.34(.05)

.29(.05)

.25(.04)

.14(.02)

Figure 10. Integrated SEMmodel for instructional strategies: mathematics

92 D. Hogan et al.

© 2013 British Educational Research Association

Page 37: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

Both of these findings, especially the former, were a surprise to us in that we expected

to find a re-assembling of the networks of pathways and number of feedback loops

from ‘higher order’ (TfU) instructional practices to ‘lower order’ (TI/DI) ones. How-

ever, in both subjects there is a linear, fully recursive sequence to instructional prac-

tice that underscores the coherent and hybridic nature of the instructional regime in

Singaporean classrooms, at least in mathematics and English, and that confirms our

arguments earlier on the basis of our CFA modelling of the four general instructional

strategies.

Table 18. Extended inventory of ‘active teaching’/‘visible learning’ indicators

Criteria

Measured standard

Core 2 data source

Secondary 3

mathematics

Secondary 3

English

Checking for prior knowledge Moderate Moderate Panel 3

Communicating learning goals Low Low Panels 2, 3.

Communicating performance standards Low Low Panels 2, 3, 5

Exemplars of successful performance Low Low Panel 3

Monitoring student learning Low Low Panels 2, 3

Feedback Low Low Panels 2, 3

Learning support/scaffolding Low Low Panel 3

Knowledge representation Low Low Panels 2, 3

Focus on metacognitive knowledge Low Low Panels 2, 3, 5

Epistemic clarity of knowledge work Low Low Panels 2, 3, 5

Open questions/extended responses Low Low Panels 2, 3

Understanding talk Low Low Panels 2, 3

Explicit instruction in norms regulating

whole class discussion and group work

Low Low Panel 3

Note: All values represent unstandardized es mates significant at p<.01, *p<.05. In brackets are standard errors.Double-headed curved arrows denotecorrelated latent disturbances.

Note. EXPR=Exam prepara on; TXBF=Textbook Focus; WKSF=Worksheet Focus; DRILL=Drill; MEM=Memoriza on; SNC=Structure & Clarity; MLT=Maximumlearning me; REV=Revision; FOP=Frequency of prac ce; FOQ=Frequency of ques oning; CFB=Collec ve feedback; PFB=Personal feedback; CNI=Curiosity &

interests; FLT=Flexible teaching; MSL=Monitoring student learning; CLGPS=Communica ng learning goals and performance standards; QTQ=Quality ofques oning; WCD=Whole class discussion; CGW=Collabora ve group work; SCF=Teacher scaffolding; FOL=Focus on Learning; CRLS=Co-regulated Learning

MSL

CFB

FOLCLGPS

SCFQTQFLT

CNI

PFB

WCD CGW

CRLS

EXPR REV

SNC

MLT

FOQ

FOPTXBF

DRILL

WKSF

MEM

.18(.04).54(.04)

.42(.04).14(.04)

.20(.05)

.37(.04).24(.04)

.19(.06).26(.05)

.14(.04)

.26(.08)

.24(.08).12(.03)*.14(.06

.69(.05)

.48(.04)

.25(.04)

.21(.03)

.31(.04)

.52(.04)

.26(.06)

.17(.04) .38(.06)

.31(.06)

.30(.06)

.77(.04)

*.10(.04) .51(.06)

.27(.06)

.11(.03)

.28(.06)

.40(.05).29(.06)

.32(.05)

.36(.05)

.26(.04)

.65(.04)

.43(.04)

.43(.04)

.47(.07)

.96(.02)

.49(.06).36(.06)

.13(.04)

.38(.07)

.30(.07)

.22(.05)

.12(.02)

.30(.06)

.34(.04)

.25(.05)

.15(.05)

.15(.04)

.20(.03)

.16(.03)

.19(.05)

.25(.07)

Figure 11. Integrated SEMmodel for instructional strategies: English

Assessment and the logic of instructional practice 93

© 2013 British Educational Research Association

Page 38: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

In broad terms, the structure goes something like this: traditional instruction

provides the foundation of the instructional order, direct instruction builds on TI

practices and extends and refines the instructional repertoire, while TfU/CRLS prac-

tices build on TI and DI practices and extend the instructional repertoire even further

in ways that focus on developing student understanding and student directed learn-

ing. Instructional practices in Singaporean classrooms then, on this data, cannot be

considered either Eastern or Western, but a coherent combination of both. However,

they combine in a particular way, as ensembles of practices grouped by broad instruc-

tional categories, rather than as new or reassembled configurations of instructional

practices that more or less ignore the instructional categories that we had imposed on

them for theoretical reasons.

What then ties or links the four instructional groupings together in an orderly chain

of instructional practice? A close review of Figures 10 and 11 indicates that four

instructional practices play pivotal roles in chaining the instructional groupings

together: two TI practice (exam preparation and textbook focus) and two DI prac-

tices (structure and clarity, and revision).

Of the four, exam preparation is the most important. In both subjects, exam prepa-

ration is highly generative both directly and indirectly, reaching well beyond its own

close family of TI practices into direct instruction and TfU practices. In mathematics,

there are nine separate direct pathways leading from exam preparation to DI and TfU

practices, and numerous indirect paths that link exam preparation, on the one hand,

to all of the remaining instructional practices, on the other. In English, exam prepara-

tion generates eight pathways to other instructional practices spread across both DI

and TfU. The aggregate indirect effect of exam preparation on focus on learning in

mathematics is .38(.02) (p < .01), and in English, .35(.02) (p < .01). The aggregate

indirect effect of exam preparation on CRLS in mathematics is .38(.02) (p < .01)

and in English, .34(.02) (p < .01).These are substantial effects, attesting to the strik-

ing institutional authority of the assessment system over instructional methods gener-

ally, not just traditional and direct instruction, and helping to account for the

remarkably isomorphic structure of instructional practice in mathematics and English

when we might well have expected quite distinctive subject-specific instructional

regimes to be generated by fundamental differences in the disciplinarity of the two

subjects.

Of course, as we have indicated elsewhere, there are other factors in play as well

that account for this instructional isomorphism, particularly the pervasive cultural

influence of a vernacular ‘folk pedagogy’ on teacher conceptions of teaching and

learning (Hogan et al., 2011, 2012d; Hogan, 2012). But there is little doubt that the

institutional and cultural authority of the assessment system, operating via exam

preparation, is exceptionally important, and that the very considerable alignment of

instruction, the curriculum and the assessment system in Singapore is in large part a

direct function of the pivotal role of exam preparation in shaping the overall structure

of the instructional system as a whole. As we indicated at the beginning of the article,

this is hardly new news to Singaporean teachers, parents, students—or researchers.

Indeed, as we saw earlier, Kramer-Dahl (2008) and others have long emphasized the

limits the assessment regime imposes on instructional practice, even when there

is a strongly reformist English syllabus in place nationally to support instructional

94 D. Hogan et al.

© 2013 British Educational Research Association

Page 39: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

innovation. Similarly, teachers we interviewed in 2010 as part of our larger research

program express remarkably similar constructions of the challenges they face in trying

to reconcile the demands of the TLLM initiative and the imperatives of high stakes

assessment. The following quotations come from two different Secondary 3 teachers:

T2: I find that for mathematics, using TLLM is a bit…How do I say, a bit tough? Because

I guess at the end of the day, what students want to see, what parents want to see, what the

school wants to see is the O level grades. So how does doing like, things like the portfolio

even, how does it value-add to what the student will gain upon their graduation. Yes, so

I’m a bit apprehensive about this TLLM, actually. (Secondary 3 mathematics teacher)

T3: I think it’s actually very difficult. Something has to give. And if the focus is so much

on the… [Sigh]. So if indeed we really want to go full steam ahead with TLLM, exams

unfortunately, have to take a backseat. But unfortunately the problem is that, the way Sin-

gaporean boys are, you know, when it’s through discovery learning, when it’s through

character development, what happens is, they may enjoy the process but enjoying the pro-

cess may not actually translate to learning the skills or motivating yourselves to do the very

best that you can. It’s like what happened years ago we had the whole concept of language

learning with the communicative approach? And it fell flat on the face. (Secondary 3

English teacher)

These teacher accounts, along with Kramer-Dahl’s account (mentioned earlier) of

the constraints imposed by the assessment system on the willingness of teachers to

implement national reform initiatives in English, is supported by evidence from a sur-

vey of over 2000 teachers we conducted in 2010. We asked teachers to identify the

impact that 21 influences had had on their teaching in Singapore—what we termed

the subjective logic of instruction.On a five-point Likert scale, ‘the ability of students’

came in first with a mean score of 4.02, followed by ‘your skills as a teacher’ (4.00),

‘coverage of the curriculum/department’s scheme of work’ (3.90) and ‘the national

high stakes assessment system’ (.3.86). We also asked teachers what influence they

thought the national high stakes assessment system had on the instructional practices

of teachers in general. We report the results in Table 21. Clearly teachers think it has

a very substantial influence, and that, in addition, it strongly influences the willing-

ness and opportunity of teachers to engage in instructional innovation.

In short, for teachers, institutional context matters—and matters a great deal. Our

strong suspicion then is that in order to explain the stability—indeed, intractability—of instructional practice in Singapore we need to recognize the persistent institutional

Table 19. Task design, student agency and knowledge building pedagogy

Criteria

Measured standard

Core 2 data sourceSecondary 3 mathematics Secondary 3 English

Epistemic agency Low Low Panels 2, 3, 5

Cognitive agency Low Very low Panels 2, 3, 5

Metacognitive agency Low Low Panels 2, 3, 5

Discursive agency Very low Very low Panels 2, 3

Assessment and the logic of instructional practice 95

© 2013 British Educational Research Association

Page 40: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

grip that the national assessment system in Singapore has over classroom practice.

While TLLM invited teachers to change their instructional practices and classroom

culture, it did not alter the national high stakes assessment system in a way that might

have increased the willingness of teachers to be more adventurous and innovative in

their classroom practices in the way TLLM hoped for. To be sure TLLM gave schools

and teachers permission to devote up to 20% of curriculum time to their own curricu-

lum preferences or interests (the so called ‘White Space’ initiative). But the anecdotal

evidence we have is that the ‘White Space’ curriculum time was quickly colonized by

more of the same and exam preparation. It’s clearly a good thing for classroom

instruction to be aligned to the assessment system, and its clearly appropriate and

sensible of teachers to align their classroom practices to it, but there are opportunity

costs as well when the degree of alignment constrains the opportunity or willingness

of teachers to alter their classroom practices in ways sought by policy-makers in

response to urgent national priorities.

Beyond exam preparation, a second traditional instructional practice is notably

generative, and generative well beyond the somewhat artificial ‘boundaries’ of

Table 20. Goodness-of-fit statistics: integrated instructional strategies SEMmodels (Secondary 3

mathematics and English)

Goodness-of-fit Mathematics English

N 1166 1027

Chi-square/df/p-value 404.786/191/.0000 353.981/194/.0000

CFI/TLI .986/.982 .989/.986

RMSEA (90% CI) .031 (.027–.035) .028 (.024–.033)SRMR .028 .031

Table 21. Teacher beliefs about the influence of national high stakes assessment on teaching

practices

Mean (1–5) SD Factor loading

Pedagogical effect of national high

stakes assessment system (alpha =.836)3.64 0.86

The national high stakes assessment system……has a very large influence on how teachers teach. 4.11 0.72 .670

…limits the willingness of teachers to try new instructional

or assessment practices.

3.64 0.92 .846

…limits the opportunity for teachers to try new instructional

or assessment practices.

3.63 0.90 .880

…pushes teachers to teach in ways contrary to their

professional beliefs.

3.46 0.90 .760

…compromises the quality of teaching. 3.36 0.88 .717

Not included in scale:

…maintains high standards of teaching and

learning

2.58 0.85

Source: Panel 2 Teacher Survey.

96 D. Hogan et al.

© 2013 British Educational Research Association

Page 41: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

traditional instruction, especially in mathematics—a focus on textbooks. In mathe-

matics, for example, there are three strong pathways from textbook focus to other TI

practices and one relatively weak one: worksheets and workbooks (.46), memoriza-

tion (.46), drill (.33) and structure and clarity (.12). But, in addition, there are three

pathways, generally weaker but still statistically significant, from textbook focus to DI

practices: frequency of practice (.27), frequency of questions (.13) and maximum

learning time (.27). These pathways all make theoretical sense, especially in mathe-

matics, where teachers tend to rely on textbooks more than they do in other subjects.

In English, textbook focus is less generative, with no pathways to drill, structure and

clarity or maximum learning time, but modest pathways to worksheets and work-

books (.24), frequency of questions (.14) and frequency of practice (.12). The aggre-

gate indirect effect of textbook focus on focus on learning in mathematics is .14(.02)

(p < .01) and in English, .03(.01) (p < .01). The aggregate indirect effect of textbook

focus on CRLS in mathematics is .14(.02) (p < .01), and in English, .02(.01)

(p < .01). Yet a third of TI practice, focus on worksheets and workbooks, had much

smaller indirect effects on the two outcome measures. The aggregate indirect effect of

worksheet focus on learning in mathematics is .02(.01) (p < .01) and in English, .04

(.01) (p < .01). The aggregate indirect effect of worksheet focus on CRLS in mathe-

matics is .02(.01) (p < .01) and in English, .04(.01) (p < .01).

The two remaining gateway instructional practices—structure and clarity and revi-

sion—belong to the DI family of instructional practices. The practice that John Hattie

closely identifies with visible learning, structure and clarity, has multiple pathways

into other DI practices and, in mathematics, into four TfU practices: collective feed-

back (.27), personal feedback (.34), curiosity and interest (.43) and communication

of learning goals and performance standards (.22). In English, it generates pathways

into four TfU practices: collective feedback (.49), personal feedback (.38), curiosity

and interest (.38) and communication of learning goals and performance standards

(.11). The aggregate indirect effect of structure and clarity on focus on learning in

mathematics is .37(.02) (p < .01), and in English, .37(.03) (p < .01). The aggregate

indirect effect of structure and clarity on CRLS in mathematics is .37(.02) (p < .01)

and in English, .35(.03) (p < .01). Revision is somewhat less generative than struc-

ture and clarity, but it has pathways into two other DI practices—frequency of prac-

tice (.16 and .26) and frequency of questions (.17 and .19)—and into three TfU

practices in both mathematics and English: flexible teaching (.37 and .30), collective

feedback (.33, .36) and personal feedback (.29, .30). The aggregate indirect effect of

revision on focus on learning in mathematics is .28(.03) (p < .01), and in English, .30

(.05) (p < .01). The aggregate indirect effect of revision on CRLS in mathematics is

.26(.03) (p < .01) and in English, .28(.04) (p < .01).

One further comment. The two proto-cognitive practices, focus on memorization

and focus on learning, occupy very different positions in Singapore’s overall instruc-

tional scheme. Focus on memorization, although it continues to be strongly net-

worked backwards to all other TI practices in the integrated model, either directly or

indirectly, remains determinedly non-generative in its own right (in mathematics)

with no pathways leading from it towards either DI or TfU practices. In effect, focus

on memorization in mathematics is as much an instructional cul de sac in the inte-

grated model as it is in the DI or the TI/DI models we discussed earlier, although in

Assessment and the logic of instructional practice 97

© 2013 British Educational Research Association

Page 42: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

English, focus on memorization generates two small to modest pathways, one to revi-

sion (.21) and one to focus on practice (.15). From these practices there are slight

indirect pathways to focus on learning, but nothing strong enough to suggest that Sin-

gaporean teachers are committed to a putative East Asian nexus between memoriza-

tion and learning, at least in mathematics and English. TfU’s focus on learning, on

the other hand, remains strongly networked to other TfU practices (with three path-

ways leading into it from other TfU practices), remains an important instructional

practice in its own right as well as generating a pathway to CRLS in both subjects.

Conclusion

In the course of this article we reported the relative importance and interrelationships

between four theoretically specified arrays of instructional strategies: traditional

instruction, direct instruction, teaching for understanding, and co-regulated learning

strategies. Specifically, we reported that traditional and direct instruction are the two

most common modalities of instructional strategy in Secondary 3 English and mathe-

matics, followed by teaching for understanding and co-regulated learning strategies.

Second, we argued that the relative weighting and overall structure of instructional

practice suggests a pervasive performative orientation to pedagogy generally in Singa-

pore. Third, we argued that instructional practice in Singapore is characterized by

hybridity rather than loyalty to an ‘East Asian’ or ‘Western’ model of pedagogy.

Finally, we used SEM modelling to establish that the hybridic nature of instructional

practice in Singapore is not institutionally innocent, but instead reflects the over-

whelming institutional authority of the assessment system over the pattern of instruc-

tional practice in Singaporean classrooms and that the alignment of instructional

practice and assessment underscores the overall performative orientation of Singa-

pore’s pedagogical framework. This arrangement, as comparative education research

suggests, is typical of systems characterized by high stakes assessment systems, and in

turn suggests that what matters most in shaping the pattern of instructional practice,

is not so much regional identities (East Asian, Western), but the broader institutional

and cultural ordering of the instructional regime.

At this point we want to note three further sets of findings that our larger research

program has generated that bear on our overall judgment of Singapore’s pedagogical

regime. The first is that using multilevel structural equation models we found that

instructional practices in Singapore that focus on procedural skills and functional

forms of cognitive activity are far better able to predict student achievement outcomes

than instructional practices that research suggest are likely to generate conceptual

understanding and related outcomes (Hogan et al., 2012d, e). Critically, we found

this pattern of relationships across a broad range of instructional practices, including

instructional tasks, classroom talk, classroom organization and instructional strate-

gies. Importantly, these findings are contrary to the Ministry of Education (MOE)

policy priorities. Our research does not indicate that teachers do not employ or value

instructional practices that focus on conceptual development or metacognitive self-

regulation or knowledge building, although their mean scores are substantially lower.

Rather, our results indicate that these more conceptually orientated instructional

practices do not have a strong impact on student achievement, given the nature of the

98 D. Hogan et al.

© 2013 British Educational Research Association

Page 43: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

current assessment system. A different assessment regime—one including different kinds

of assessment tasks, for example—would in all likelihood generate a different pattern

of relationships and do so in line with current policy priorities. So while the current

assessment system has improved teaching and learning standards over the years, it

now appears that the current assessment regime inhibits or constrains the willingness

and opportunity of teachers to change their instructional practices in line with current

policy priorities that favours what Singaporeans loosely (and misleadingly) view as

‘learner-centered’ pedagogy. In effect, the current assessment regime incentivizes and

rewards teachers to teach (and students to learn) in ways that maximize assessment perfor-

mance rather than the kinds of teaching and learning called for in national policy documents

and generally associated with teaching for understanding frameworks. This suggests that

current policy settings are not internally consistent, as we indicated above.

The second additional finding we want to report now is that we found that prior

achievement was consistently the single strongest predictor of student achievement.

However, we also identified the existence of very large direct and indirect social class

effects on prior achievement (PSLE) and student allocation to curriculum streams

and on Secondary 3 student achievement at the classroom level. Social class effects at

the individual (L1) level were much smaller, although larger in the case of English

than mathematics, a result that we think underscores the importance of linguistic,

social and cultural capital at the family level. But although social class effects at the

individual level within classes were relatively small, there can be no mistaking the sig-

nificance of the impact of social class on Secondary 3 student achievement at the

aggregate classroom level. In Singapore, as in other systems committed to homoge-

neous grouping, it appears to matter less what particular family students come from

then who they go to class (and school) with. To put it in technical terms: institutional

and composition effects matter, and matter a great deal. This is consistent with inter-

national research, but the relationship between social class and educational inequality

in Singapore is exaggerated by institutional rules and organizational arrangements

(above all, the streaming system) by virtue of the very tight relationship with social

class and streaming in Singapore that we (and others) have quantified. In effect, in

Singapore, streaming compounds and inflates the influence of social class on student

achievement at the classroom and school level. (By contrast, systems with lower levels

of aggregation are typically characterized by higher social class effects at the individ-

ual student level).

These particular findings have very substantial policy implications for the improve-

ment of teaching and learning across the system and the design of interventions

intended to ameliorate the effects of social inequality in education. In particular, they

suggest that the improvement of teaching and learning will depend substantially on a

different assessment regime, including, in particular, alteration in the nature of the

assessment tasks in the high stakes assessment system, and a different kind of curricu-

lum framework to give appropriate guidance to the enacted as well as the prescribed

curriculum. In addition, our evidence suggests that if the system wishes to reduce

social class differences in student performance, it will need to address sources

(including impacts on student motivation and identity formation and access to

human, social and cultural capital at the classroom level) of inequality that are a func-

Assessment and the logic of instructional practice 99

© 2013 British Educational Research Association

Page 44: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

tion of organizational arrangements and policy settings as well as those that arise at

the family and individual level.

Finally, again using structural equation procedures, we found that teacher effects,

independent of instructional or teaching effects, are minimal. Our measures of

teacher effects included pre-service education, participation in professional develop-

ment activities, participation in in situ professional learning communities (profes-

sional collaboration and reflective dialogue), teaching experience, learning priorities

(‘problem-solving efficiency’ versus ‘learning to learn’), conceptions of teaching and

learning (‘conventional’, ‘constructivist’), self-efficacy beliefs (with respect to class-

room management, student motivation and instruction), and ability and effort attri-

butions regarding student achievement. Indeed, in both mathematics and English,

the only statistically significant pathway we could identify was from one of the learn-

ing priorities measures (‘problem-solving efficiency’) to our student achievement

measures. We interpret this finding to reflect (and reinforce) the performative orien-

tation of Singapore’s pedagogy. In addition, we found that the pattern of relationships

between participation in professional development and professional learning commu-

nity activities, on the one hand, and instructional beliefs and student outcomes, on

the other, suggest that professional development and professional learning commu-

nities reinforce rather than challenges the dominant performative orientation of peda-

gogical practice in Singapore. This in turn suggests some tension between MOE

policy priorities (in this case, a commitment to learning to learn) and the overall struc-

turing of teacher beliefs and participation in professional development and profes-

sional learning communities.

Of course, these limitations of Singapore’s instructional regime need to be placed

in the broader context of a highly successful system that by any measure has gener-

ated an extraordinary record of achievement over the past two or three decades so

that it is now widely recognized as one of the leading educational systems in the

world. But notwithstanding this record of achievement, our judgment now is that

while the performative and hybridic character of the system has served Singapore

well, it is far from clear that the current framework can support successful, substantial

and sustainable innovation in classroom practice. From a normative perspective, the

current system is suboptimal in a number of respects—the small mean scores for TfU

practices as a whole, particularly whole class discussion, collaborative group work

and focus on learning and the absence of key pathways, including pathways from

monitoring student learning to student feedback (collective and personal). Finally—and perhaps the most concerning of all—is the apparent constraint that the assess-

ment system places on the willingness and opportunity of teachers to alter instruc-

tional practice, even when urged to do so by relevant national syllabi or by national

reform initiatives. This might suggest that the sensible thing to do would be to elimi-

nate the national high stakes assessment system. We don’t agree. Rather, the chal-

lenge for the Ministry of Education, teachers and researchers is to figure out how to

transform and use the institutional authority of the assessment system as a lever for

change rather than a constraint on improvement. Our sense is that this might be

accomplished in four ways: (1) by improving the quality of the assessment tasks in the

national high stakes assessment system, and doing so in a way that prioritizes

extended, elaborated, authentic, multidimensional twenty-first century knowledge

100 D. Hogan et al.

© 2013 British Educational Research Association

Page 45: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

building tasks (including tasks that are both collaborative and ICT-mediated) that

will drive instructional improvement, given the strong proclivity of teachers to teach

to the test; (2) by introducing a school-based, professionally moderated, standards-

driven component (say initially, accounting for up to 30% of the final grade) into the

national high stakes assessment system; (3) by enhancing the professional capability,

authority and pedagogical autonomy of teachers in ways that allows them to moderate

or buffer the demands that emanate from exogenous pressures outside the classroom;

and (4) by a modest additional deregulation of the assessment market in Singapore

(some, but not all, Singaporean schools are permitted to offer the International

Baccalaureate (IB) rather than participate in the Cambridge assessment system)

granting all schools permission to choose which national assessment regime to join in

a semi-deregulated assessment market—Cambridge, IB, or a hybrid that combines

national plus school-based components—depending on what pedagogical priorities

schools establish for themselves in the spirit of devolved pedagogical authority that

TLLM supported in 2004/2005.

For purists, the proposal to use the assessment system to support instructional

improvement might well appear misconceived, or at least paradoxical, given the well

documented negative effects national high stakes assessment systems can have. And

there is some evidence that the one experiment in school based assessment in Singa-

pore we are familiar with, an experiment using ‘science practical assessments’ as a form

of school based assessment lacked the capacity to drive significant pedagogical change

(in this case, inquiry science) without a considerable investment in teacher preparation

and professional development and other collateral changes. Rather, it cornered teach-

ers into assessing that which could bemost easily observed in laboratories and exposed

fears from a range of actors about a lack of content knowledge and assessment literacy

(Towndrow & Tan, 2006; Towndrow, 2008; Towndrow et al., 2010). Consequently,

we conclude that school-based assessment by itself will not achieve instructional

improvement. Rather, instructional improvement is primarily likely to depend on

transforming the nature of high stakes assessment tasks, backed up by school-based

high stakes assessments and significant investments in building teacher capacity.With-

out these changes in the assessment system, we can see no other solution in sight that

has the capacity to be as effective, sustainable, scaleable and, critically, politically man-

ageable in the Singaporean context. In Singapore (as elsewhere), instructional legiti-

macy depends on formal inclusion in the assessed curriculum. More broadly,

instructional systems, like organizational systems generally, are prone to isomorphism

with their institutional environments (Powell & DiMaggio, 1991; Scott, 1995; Hogan

et al., 2008b). Unless assessment tasks that focus on complex knowledge work are

properly integrated into the assessed curriculum, students, parents or teachers will not

take complex knowledge instruction and related assessment tasks seriously. If this were

accomplished, the tight coupling (or alignment) of instruction and summative assess-

ment (‘teaching to the test’) in Singapore becomes a pedagogical strength rather than a

constraint on sustainable pedagogical innovation. In effect, appropriately constructed

high stakes assessments might, other things being equal, leverage desirable instruc-

tional innovation and pedagogical realignments in classrooms across the system in a

sustainable way at scale and at verymodest financial (and, importantly, political) cost.

Assessment and the logic of instructional practice 101

© 2013 British Educational Research Association

Page 46: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

NOTE

1 However, a very large classroom observation data set from a related project we are conducting suggests a farbleaker picture on this front. See Hogan et al. (2012d).

References

Alexander, R. (2001) Culture and pedagogy: International comparisons in primary education (Oxford,

Blackwell).

Alexander, R. (2008) Essays on pedagogy (London, Routledge).

Alexander, R. (2012) Improving oracy and classroom talk in English schools: Achievements and

challenges, paper presented to the DfE Seminar on Oracy, the National Curriculum and Educa-

tional Standards, 20 Febuary, Department of Education, UK.

Assor, A., Kaplan, H., Kanat-Maymon, Y. & Guy, R. (2005) Directly controlling teacher behaviors

as predictors of poor motivation and engagement in girls and boys: The role of anger and anxi-

ety, Learning and Instruction, 15(5), 397–413.Bandalos, D. & Finney, S. (2001) Item parceling issues in structural equation modeling, in:

G. Marcoulides (Ed.) New developments and techniques in structural equation modeling (Mahwah,

NJ, Lawrence Erlbaum), 269–296.Barnes, D. (1992) From communication to curriculum (Portsmouth, NH, Boynton/Cook-Heinemann).

Barnes, D. (2008) Exploratory talk for learning, in: N. Mercer & S. Hodgkinson (Eds) Exploring talk

in school (London, Sage), 1–16.Biggs, J. & Walkins, D. (2001) Teaching the Chinese learner: Psychological and pedagogical perspectives

(Hong Kong, Comparative Education Research Centre).

Black, P., Harrison, C., Lee, C., Marshall, B. & Wiliam, D. (2003) Assessment for learning: Putting it

into practice (Milton Keynes, Open University Press).

Black, P. & Wiliam, D. (1998) Inside the black box: Raising standards through classroom assessment

(London, School of Education).

Bollen, K. (1989) Structural equations with latent variables (New York, John Wiley & Sons).

Brophy, J. E. (2002) Social constructivist teaching: Affordances and constraints (J. E. Brophy, Ed.)

(Amsterdam, JAI Press).

Brophy, J. E. (2004)Motivating students to learn (Mahwah, NJ, Lawrence Erblaum).

Brown, T. (2006) Confirmatory factor analysis for applied research (New York, Guilford Press).

Bruner, J. (1996) Folk pedagogy, in: J. Bruner (Ed) The culture of education (Cambridge, MA,

Harvard University Press), 44–65.Cai, J., Lin, F.-L. & Fan, L. (2004) How do Chinese learn mathematics? Some evidence-based

insights and needed directions, in: L. Fan & L. Fan (Eds) How Chinese learn mathematics:

Perspectives from insiders (Singapore, World Scientific), 535–554.Cazden, C. (1988) Classroom discourse: The language of teaching and learning (Portsmouth, NH,

Heinemann).

Chan, M. (2012) A latent variable method to constructing reliable composite variables. Unpublished

manuscript, National Institute of Education, Singapore.

Cheah, Y. M. (1998) The examination culture and its impact on literacy innovations: The case of

Singapore, Language and Education, 12(3), 192–209.Cheng, K. M. (1996) Excellence in education: Is it culture-free? Keynote paper presented at the

annual Conference of the Educational Research Association, Singapore, 22–4 November.

Coffman, D. L. & MacCullum, R. C. (2005) Using parcels to convert path analyses models into

latent variable models,Multivariate Bahvioral Research, 40(2), 235–259.Cohen, D. (1988) Teaching practice: Plus que ca change…, in: P. W. Jackson (Ed.) Contributing to

educational change: Perspectives on research and practice (Berkeley, CA, McCutchan), 27–84).(Also published in the National Center for Research on Teacher Education, Michigan: Michi-

gan State University, 88-3, September 1988.)

Darling-Hammond, L. (2008) A future worthy of teaching for America, Phi Delta Kappan, 89(10),

730–733.

102 D. Hogan et al.

© 2013 British Educational Research Association

Page 47: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

Desforges, C. (Ed.) (1995) An introduction to teaching: Psychological perspectives (Oxford, Blackwell).

Galton, M. (2007) Learning and teaching in the Primary classroom (London, Sage).

Goh, C. T. (1997) Shaping our future: Thinking schools, learning nation. Speech by Prime Minis-

ter Goh Chok Tong at the opening of the 7th International Conference on Thinking. Available

online at: http://www.moe.gov.sg/media/speeches/1997/020697.htm (accessed 6 March 2009).

Good, T. L. & Brophy, J. E. (2003) Looking in classrooms (New York, Allyn & Bacon).

Hacker, D., Dunlosky, J. & Graesser, A. (2009) Handbook of metacognition in education (London,

Routledge).

Hattie, J. (2003) Distinguishing expert teachers from novice and experienced teachers (Auckland, Austra-

lian Council for Educational Research).

Hattie, J. (2009) Visible learning. A synthesis of over 800 meta-analyses relating to achievement (London,

Routledge).

Hattie, J. (2012) Visible teaching for teachers: Maximizing impact on learning (London, Routledge).

Hodgkinson, S. &Mercer, N. (2008) Exploring talk in school (London, Sage).

Hogan, D. (2012) Culture and pedagogy in Singapore: The fate of the Teach Less Learn More policy

initiative, 2004–2010, in: S. Paris & K. Lee (Eds) Redesigning pedagog (New York, Springer).

Hogan, D., Towndrow, P. & Koh, K. (2008a) Instruction and assessment practice in Singapore, in:

E. Grigorenko, (Ed.) Assessment of abilities and competencies in the era of globalization (New York,

Springer).

Hogan, D., Towndrow, P. & Koh, K. (2008b) The logic of confidence and the social economy of

assessment reform in Singapore: A new institutionalist perspective, in: E. Grigorenko (Ed.)

Assessment of abilities and competencies in the era of globalization (New York, Springer).

Hogan, D., Towndrow, P., Abdul Rahim, R., Chan, M., Luo, S. & Sheng, Y., et al. (2011) Interim

report on pedagogical practices in Singapore in Secondary 3 mathematics and English, 2004 and 2010

(Singapore, National Institute of Education).

Hogan, D., Rahim, R., Chan, M., Kaur, B. & Towndrow, P. (2012a) Disciplinarity and the logic of

mathematical tasks in Secondary 3 mathematics lessons in Singapore, in: R. Gillies (Ed.) New

developments in cognition and instruction research (New York, Nova Science), 337–370.Hogan, D., Chan, M., Rahim, R., Towndrow, P. & Kwek, D. (2012b) Understanding classroom

talk in Secondary 3 mathematics classes in Singapore, in: B. Kaur (Ed.) Reasoning, communica-

tion and connections in mathematics: Yearbook 2012, Association of mathematics educators

(Singapore, World Scientific), 168–198.Hogan, D., Kwek, D., Towndrow, P., Rahim, R., Tan, T. K., Kaur, B., Yang, H. J. & Chan, M.

(2012c) Opaque or transparent? First reflections on visible learning in Singapore, in: S. Deng

Zongi, S. Gopinathan & C. Lee (Eds) Globalization and the Singapore curriculum: From policy to

classroom (London, Springer).

Hogan, D., Towndrow, P., Kwek, D., Rahim, R., Chan, M., Aye, K. & Teh, L.W. (2012d) Core 2

Interim Report 2 (Singapore, National Institute of Education).

Hogan, D., Chan, M. & Ridzuan, R. (2012e) The epistemic and cognitive framing of instructional

tasks in Secondary 3 mathematics and English classes in Singapore and their relationship to

student academic performance, paper presented at the Australian Association for Research in

Education Conference, Sydney, December 4.

Hu, L. & Bentler, P. (1999) Cutoff criteria for fit indexes in covariance structure analysis: Conven-

tional criteria versus new alternatives, Structural Equation Modeling, 6(1), 1–55.James, M., McCormick, R., Black, P., Carmichael, P., Drummond, M.-J., Fox, A., et al. (2007)

Improving learning how to learn: Classrooms, schools and networks (Abingdon, Routledge).

J€oreskog, K. G. (1971) Statistical analysis of sets of congeneric tests, Psychometrika, 36(2), 109–133.

Kline, R. B. (2011) Principles and practice of structural equation modeling (New York, The Guilford

Press).

Kramer-Dahl, A. (2008) Negotiating what counts as English language teaching: Official curriculum

and its enactment in two Singaporean secondary classrooms, Research Papers in Education, 23

(1), 85–107.

Assessment and the logic of instructional practice 103

© 2013 British Educational Research Association

Page 48: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

Landis, R. S., Beale, D. J. & Tesluk, P. E. (2000) A comparison of approaches to forming compos-

ite measures in structural equation modelling, Organizational Research Methods, 3(2), 186–207.Leung, F. (2001) In search of an East Asian identity in mathematics education: The legacy of an

old culture and the impact of modern technology, Educational Studies in Mathematics, 47(1),

35–41.Leung, F. (2006) Mathematics education in East Asia and the west: Does culture matter?, in:

F. Leung, K.-D. Graf & F. Lopez-Real (Eds) Mathematics education in different cultural

traditions: A comparative study of East Asia and the west (New York, Springer), 21–46.Luo, W., Hogan, D. & Paris, S. G. (2011a) Predicting Singapore students, achievement goals in

their English study: Self-construal and classroom goas structure, Learning anf Individual

Differences, 21(5), 526–535.Luo, W., Hogan, D., Paris, S. G. & Luo, Z. (2011b) Do performance goals promote learning?

A pattern analysis of Singapore students’ achievement goals, Contemporary Educational

Psychology, 36(2), 165–176.Mercer, N. (1992) Talk for teaching and learning, in: K. Norman (Ed) Thinking voices: The work of

the National Oracy Project (London, Hodder & Stoughton), 215–223.Mercer, N. & Littleton, K. (2007) Dialogue and the development of children’s thinking: A sociocultural

approach (Abingdon, Routledge).

Michaels, S., O’Connor, C., Hall, M. & Resnick, L. (2002) Accountable talk: Classroom conversation

that works (CD-ROM set) (Pittsburgh, University of Pittsburgh).

Michaels, S., O’Conner, C. & Resnick, L. (2008) Deliberative discourse idealized and realized:

Accountable talk in the classroom and civic life, Studies in the Philosophy of Education, 27(4),

283–297.Michaels, S., Sohmer, R.E. & O’Connor, M.C. (2004) Classroom discourse, in: H. Ammon,

N. Dittmar, K. Mattheier & P. Trudgill (Eds) Sociolinguistics: An international handbook of the

science of language and society (New York, Walter de Gruyter), 2351–2366.Ministry of Education (MOE) (2008) More support for school’s ‘Teach less, learn more’ initiatives,

para. 6–10. Available online at: http://www.moe.gov.sg/media/press/2008/01/more-support-

for-schools-teach.php (accessed 6 February 2010).

Mok, M., Cheng, Y.C., Moore, P. & Kennedy, P. (2006) The development and validation of the

self-directed learning scale, Journal of Applied Measurement, 7(4), 419–449.National Mathematics Advisory Panel (NMAP) (2008) Report of the task force on instructional

practices (Washington, DC, Department of Education).

National Research Council (NRC) (2000) How people learn (Washington, DC, National Academy

Press).

Newmann, F. M., Marks, H. M. & Gamoran, A. (1995) Authentic pedagogy: Standards that boost

student performance. Centre on Organization and Restructuring of Schools, Issues in Restruc-

turing Schools, Report Number 8 (Madison, WI, University of Wisconsin-Madison).

Nota, l., Soresi, S. & Zimmerman, B. (2004) Self-regulation and academic achievement and

resilience: A longitudinal study, International Journal of Educational Research, 41(3), 198–215.Nystrand, M., Gamoran, A., Kachur, R. & Prendergast, C.D. (1999) Opening dialogue: Under-

standing the language and learning in the English classroom (New York, Teachers College

Press).

Nystrand, M., Wu, L., Gamoran, A., Zeiser, S. & Long, D. (2001) Questions in time: Investigating

the structure and dynamics of unfolding classroom discourse (Albany, The National Research Center

on English Learning & Achievement, Report Series 14005, University of New York).

OECD (1999) Measuring student knowledge and skills: A new framework for assessment (Paris,

OECD).

OECD (2010) PISA 2009 key findings. What students know and can do: Student performance in reading,

mathematics and science (vol. I). Available online at: http://www.oecd.org/pisa/pisaproducts/

48852548.pdf (accessed 7 February 2012).

Paris, S. & Paris, A. (2001) Classroom applications of research on self-regulated learning,

Educational Psychologist, 36(2), 89–101.Perkins, D. (1993) Teaching for understanding, American Educator, 17(3), pp. 8, 28–35.

104 D. Hogan et al.

© 2013 British Educational Research Association

Page 49: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

Perkins, D. (1998) What is understanding?, in: M.S. Wiske (Ed) Teaching for understanding (San

Francisco, Jossey Bass), 39–57.Powell, W.W. & DiMaggio, P.J. (1991) The iron cage revisited: Institutional isomorphism and col-

lective ratinality, in: W.W. Powell & P.J. DiMaggio (Eds) The new institutionalism in organiza-

tional analysis (Chicago, University of Chicago Press), 63–82.Preacher, K. & Kelley, K. (2011) Effect size measures for mediation models: Quantitative strategies

for communicating indirect effects, Psychological Methods, 16(2), 93–115.Purdie, N. & Ellis, L. (2005) A review of the emphirical evidence identifying effective interventions and

teaching practices for students with learning difficulties in Years 4,5 and 6. Learning Processes and

Contexts Research Program (Camberwell, Victoria, Australian Council for Educational

Research). http://www.acer.edu.au/documents/Purdie-Ellis_LiteratureReview.pdf (accessed

7 December 2012).

Putnam, R., Lampert, M. & Peterson, P. (1990) Alternative perspectives on knowing mathematics

in elementary schools, Review of Research in Education, 16(1), 57–150.Rahim, R., Hogan, D. & Chan, M. (2012) The epistemic framing of mathematical tasks in Second-

ary 3 mathematics lessons in Singapore, in: B. Kaur (Ed.) Reasoning, communication and connec-

tions in mathematics: Yearbook 2012, Association of mathematics educators (Singapore, World

Scientific), 11–56.Resnick, L.B., Michaels, S. & O’Connor, C. (2010) How (well structured) talk builds the mind,

in: R. Sternberg & D. Preiss (Eds) From genes to context: New discoveries about learning from

educational research and their applications (New York, Springer), 163–194.Rittle-Johnson, B. & Alibali, M.W. (1999) Conceptual and procedural knowledge of mathematics:

Does one lead to the other?, Journal of Educational Psychology, 91(1), 175–189.Rowe, K. J. (2002) The measurement of latent and composite variables frommultiple items or indica-

tors: Applications in performance indicator systems, background paper to keynote address pre-

sented at the RoyalMelbourne Institute of Technology Statistics Seminar Series, October 11.

Rowe, K. J. (2003) The importance of teacher quality as a key determinant of students’ experiences

and outcomes of schooling, background paper to invited address presented at the 2003 ACER

Research Conference, Carlton Crest Hotel, Melbourne, 19–21 October.

Rowe, K.J. (2006) School performance: Australian state/territory comparisons of student achievements in

national and international studies (Camberwell, Victoria, Australian Council for Educational

Research).

Sass, D.A. & Smith, P.L. (2006) The effects of parceling unidimensional scales on structural

parameter estimates in structural equation modelling, Structural Equation Modeling, 13(4),

566–586.Schoenfeld, A.H. (1992) Learning to think mathematically: Problem-solving, metacognition, and

sense-making in mathematics, in: D. Grouws (Ed.) Handbook for research on mathematics

teaching and learning (New York, MacMillan), 334–370.Schraw, G. (2006) Knowledge: Structures and processes, in: P. Alexander & P. Winne (Eds)

Handbook of educational psychology (Mahwah, Lawrence Erblaum), 245–264.Scott, W.R. (1995) Institutions and organizations (Thousand Oaks, CA, Sage).

Stein, M.K., Grover, B.W. & Henningsen, M. (1996) Building student capacity for mathematical

thinking and reasoning: An analysis of mathematical tasks used in reform classrooms, American

Educational Research Journal, 33(2), 455–488.Stein, M., Smith, M., Henningsen, M. & Silver, E. (2009) Implementing standards-based mathematics

instruction: A casebook for professional development (New York, Teachers College Press).

Stephenson, M.T. & Holbert, R.L. (2003) A Monte Carlo simulation of observable- versus latent-

variable structural equation modeling techniques, Communication Research, 30(3), 332–354.Towndrow, P.A. (2008) Critical reflective practice as a pivot in transforming science education:

A report of teacher–researcher collaborative interactions in response to assessment reforms,

International Journal of Science Education, 30(7), 903–922.Towndrow, P. A. & Tan, A. L. (2006) Transforming science practical assessment practices in Sin-

gapore through innovative departmental planning, paper presented at the International Science

Education Conference 2006, ‘Science education: What works?’, November, Singapore.

Assessment and the logic of instructional practice 105

© 2013 British Educational Research Association

Page 50: Assessment and the logic of instructional practice in Secondary 3 English and mathematics classrooms in Singapore

Towndrow, P.A., Tan, A.L., Yung, B.H.W. & Cohen, L. (2010) Science teachers’ professional

development and changes in science practical assessment practices: What are the issues?,

Research in Science Education, 40(2), 117–132.Urdan, T. & Midgley, C. (2001) Academic self-handicapping: What we know, what more there is

to learn, Educational Psychology Review, 13(2), 115–138.Wiggins, G. P. &McTighe, J. (2005) Understanding by design (2nd ed.). (Alexandria, VA: ASCD).

Wiliam, D. (2007) Keeping learning on track: Classroom learning and the regulation of assessment,

in: K. Lester (Ed.) Second handbook of research on mathematics teaching and learning (Charlotte,

NC: Information Age Publishing), 1053–1098.Wolters, C. (2004) Advancing achievement goal theory: Using goal structure and goal orientation

to predict students’ motivation, cognition, and achievement, Journal of Educational Psychology,

96(2), 236–250.Wong, N.-Y. (2004) The CHC learner’s phenomenon: Its implications on mathematics education,

in: L. Fan & L. Fan (Ed.) How Chinese learn mathematics: Perspectives from insiders (Singapore,

World Scientific Publishing), 503–534.Zimmerman, B.J. & Schunk, D.H. (2011) Handbook of self-regulation of learning and performance

(London, Routledge).

106 D. Hogan et al.

© 2013 British Educational Research Association