25
The effects of teacher mathematics knowledge and pedagogy on student achievement in rural Guatemala Jeffery H. Marshall M. Alejandra Sorto Published online: 22 February 2012 Ó Springer Science+Business Media B.V. 2012 Abstract Why are some teachers more effective than others? The importance of understanding the interplay between teacher preparation, pedagogy and student achievement has motivated a new line of research focusing on teacher knowledge. This study analyses the effects of teacher mathematics knowledge on student achievement using longitudinal data from rural Guatemalan primary schools. After presenting a conceptual framework for linking the work of the teacher with student learning in mathematics together with an overview of the different forms of teacher knowledge, the paper introduces the Guatemalan context and the analytical framework including the sample, data and methods. Overall, the results provide some empirical support for a widely held, if infrequently tested, belief in mathe- matics education: effective teachers have different kinds of mathematical knowl- edge. The results also suggest specific mechanisms by which effective teachers can make substantial impacts on student learning, even in extremely poor contexts. Keywords Student mathematics achievement Teacher mathematics knowledge Teaching pedagogy Guatemala Resume ´ Effets des connaissances mathe ´matiques et de la pe ´dagogie des enseig- nants sur les re ´sultats scolaires dans le Guatemala rural – Pourquoi certains en- seignants sont plus efficaces que d’autres ? L’importance de cerner l’interaction entre pre ´paration de l’enseignant, pe ´dagogie et re ´sultats scolaires a motive ´ un nouvel axe de recherche qui se concentre sur les connaissances de l’enseignant. La pre ´sente e ´tude analyse les conse ´quences des connaissances mathe ´matiques des J. H. Marshall (&) 4610 Ironstone Lane, West Lafayette, IN 47906, USA e-mail: [email protected] M. A. Sorto 601 University Drive, San Marcos, TX 78666, USA e-mail: [email protected] 123 Int Rev Educ (2012) 58:173–197 DOI 10.1007/s11159-012-9276-6

The effects of teacher mathematics knowledge and pedagogy on student achievement in rural Guatemala

Embed Size (px)

Citation preview

The effects of teacher mathematics knowledgeand pedagogy on student achievement in ruralGuatemala

Jeffery H. Marshall • M. Alejandra Sorto

Published online: 22 February 2012

� Springer Science+Business Media B.V. 2012

Abstract Why are some teachers more effective than others? The importance of

understanding the interplay between teacher preparation, pedagogy and student

achievement has motivated a new line of research focusing on teacher knowledge.

This study analyses the effects of teacher mathematics knowledge on student

achievement using longitudinal data from rural Guatemalan primary schools. After

presenting a conceptual framework for linking the work of the teacher with student

learning in mathematics together with an overview of the different forms of teacher

knowledge, the paper introduces the Guatemalan context and the analytical

framework including the sample, data and methods. Overall, the results provide

some empirical support for a widely held, if infrequently tested, belief in mathe-

matics education: effective teachers have different kinds of mathematical knowl-

edge. The results also suggest specific mechanisms by which effective teachers can

make substantial impacts on student learning, even in extremely poor contexts.

Keywords Student mathematics achievement � Teacher mathematics knowledge �Teaching pedagogy � Guatemala

Resume Effets des connaissances mathematiques et de la pedagogie des enseig-

nants sur les resultats scolaires dans le Guatemala rural – Pourquoi certains en-

seignants sont plus efficaces que d’autres ? L’importance de cerner l’interaction

entre preparation de l’enseignant, pedagogie et resultats scolaires a motive un

nouvel axe de recherche qui se concentre sur les connaissances de l’enseignant. La

presente etude analyse les consequences des connaissances mathematiques des

J. H. Marshall (&)

4610 Ironstone Lane, West Lafayette, IN 47906, USA

e-mail: [email protected]

M. A. Sorto

601 University Drive, San Marcos, TX 78666, USA

e-mail: [email protected]

123

Int Rev Educ (2012) 58:173–197

DOI 10.1007/s11159-012-9276-6

enseignants sur les performances des eleves, en exploitant les donnees longitudi-

nales issues d’ecoles primaires du Guatemala rural. Les auteurs presentent un cadre

conceptuel qui relie le travail de l’enseignant a l’apprentissage des mathematiques

par les eleves, ainsi qu’une vue d’ensemble des diverses formes de savoir de

l’enseignant. Ils exposent ensuite le contexte guatemalteque et le cadre analytique

comportant l’echantillon, les donnees et les methodes. Globalement, les resultats

apportent un certain soutien empirique a la croyance largement repandue mais guere

verifiee sur l’enseignement des mathematiques : les enseignants efficaces possedent

des formes differentes de connaissances mathematiques. Les resultats induisent en

outre des mecanismes specifiques permettant aux enseignants efficaces d’exercer

une influence sensible sur l’apprentissage des eleves, meme dans les contextes

d’extreme pauvrete.

Zusammenfassung Die Auswirkungen mathematischer und padagogischer

Fahigkeiten von Lehrkraften auf die Leistungen ihrer Schulerinnen und Schuler im

landlichen Guatemala – Warum sind manche Lehrerinnen und Lehrer erfolgreicher

als andere? Aus dem dringenden Bedurfnis, die Wechselwirkungen zwischen

Lehrerausbildung, Padagogik und den Lernerfolgen der Schulerinnen und Schuler

zu verstehen, ist eine neue Forschungsrichtung erwachsen, die sich vor allem mit

dem Fachwissen der Lehrkrafte befasst. Anhand von Langsschnittdaten landlicher

Primarschulen in Guatemala wird in dieser Studie analysiert, wie sich das Wissen

der Lehrkrafte im Fach Mathematik auf die Leistungen ihrer Schulerinnen und

Schuler auswirken. Auf die Darstellung eines Rahmenkonzepts, das die Verknup-

fung zwischen der Arbeit der Lehrkraft und den Lernerfolgen der Schulerinnen und

Schuler im Fach Mathematik herstellt, und eines Uberblicks uber die verschiedenen

Arten von Lehrerwissen folgt eine Einfuhrung in den guatemaltekischen Kontext

und den Analyserahmen, einschließlich Stichprobe, Daten und Methoden. Insge-

samt liefern die Ergebnisse einen empirischen Beleg fur eine haufig anzutreffende,

jedoch selten uberprufte Ansicht im Bereich der mathematischen Bildung: erfol-

greiche Lehrkrafte verfugen uber verschiedene Arten von mathematischen Kennt-

nissen. Die Ergebnisse deuten außerdem darauf hin, dass erfolgreiche Lehrkrafte die

Lernleistungen von Schulerinnen und Schulern, selbst in einem Kontext extremer

Armut, durch bestimmte Mechanismen erheblich beeinflussen konnen.

Resumen Efectos de conocimientos en matematicas y pedagogıa de los docentes

sobre el rendimiento estudiantil en las zonas rurales de Guatemala – >Por que

algunos docentes son mas efectivos que otros? La importancia de entender la in-

teraccion entre preparacion docente, pedagogıa y rendimiento estudiantil ha dado

lugar a una nueva lınea de investigacion enfocada en los conocimientos del docente.

En este estudio, usando datos longitudinales de escuelas primarias en zonas rurales

de Guatemala, se analizan los efectos que tienen los conocimientos en matematicas

del docente sobre el rendimiento estudiantil. Luego de presentar un marco con-

ceptual para enlazar el trabajo del docente con el aprendizaje de matematicas de los

estudiantes, junto con una vision general de las diferentes formas de conocimiento

docente, este trabajo introduce el contexto Guatemalteco y el marco analıtico, in-

cluyendo muestra, datos y metodos. En general, los resultados proveen algun apoyo

174 J. H. Marshall, M. A. Sorto

123

empırico a la creencia generalizada, aunque poco comprobada, en cuanto a

ensenanza de las matematicas: los docentes efectivos tienen diferentes clases de

conocimientos matematicos. Los resultados tambien indican mecanismos especıfi-

cos con los cuales los docentes efectivos pueden producir un impacto importante en

el aprendizaje de los estudiantes, incluso en contextos de extrema pobreza.

Introduction

It makes intuitive sense that a teacher’s mathematics knowledge affects his or her

ability to teach mathematics and, by extension, his or her students’ achievement.

But considering the results of more than forty years of statistical analyses of student

test scores there is surprisingly little evidence to make this case empirically (Wayne

and Youngs 2003). Most studies have had to make do with proxies of the teacher’s

knowledge based on mathematics coursework, certification and education levels.

These are poor substitutes for actual knowledge, but less intrusive data to collect

compared with teacher tests. The exceptions have, historically, come mainly from

developing countries where a handful of studies show a positive correlation between

the teacher’s knowledge of mathematics and student achievement (Harbison and

Hanushek 1992; Mullens et al. 1996; Marshall and White 2001).

The results from these studies make a general case for raising teacher content

knowledge through more intensive pre-service preparation or in-service training.

But a number of important questions remain. For example, how much exposure to

higher-level mathematics content do teachers need to be effective, including non-

mathematics specialists or general education majors who often work in primary

schools? Also, to what degree is the empirical link between mathematics content

knowledge and student achievement a product of an underlying, specialised

The effects of teacher mathematics knowledge and pedagogy 175

123

teaching knowledge domain (Shulman 1986)? And how does teachers’ knowledge

of mathematics influence their pedagogical choices (i.e. how much time to spend on

which activities) in the classroom?

The importance of understanding this interplay between teacher preparation,

pedagogy and student achievement has motivated a new line of research focusing on

teacher knowledge. A lot of the work builds on the pedagogical content knowledge

(or PCK) concept introduced by Lee Shulman (Shulman 1986). PCK goes beyond

basic content knowledge and emphasises subject matter knowledge for teaching,

which can include the teacher’s knowledge of how students learn as well as different

ways of teaching specific content. In recent years these ideas have been further

extended in mathematics through the work of Deborah Ball, Hyman Bass and

Heather Hill (Ball and Bass 2000; Hill and Ball 2004; Hill et al. 2008a). As a result

there is a growing research base analysing different forms of teacher knowledge

using large-sample surveys of student achievement. For example, Hill et al. (2005)

and Jonah E. Rockoff et al. (2008) found that the teacher’s ‘‘mathematics

knowledge for teaching’’ (or MKT) predicts student knowledge growth in

elementary and middle schools in the United States. Jurgen Baumert et al. (2010)

found that a measure of PCK is a significant predictor of student mathematics

learning gains in German secondary schools. Marshall et al. (2009) have shown

significant results linking measures of teacher PCK and content knowledge with

student achievement in Cambodian primary schools.

Our study continues this line of inquiry in analysing the effects of teacher

mathematics knowledge on student achievement using longitudinal data from rural

Guatemalan primary schools. The teachers in our sample answered mathematics test

questions at primary and middle school levels, and completed several activities

designed to measure their specialised content knowledge for primary school

mathematics teaching. This information is augmented with observational data that

break down lessons into a series of time segments for activities such as seatwork1

and recitation.2 The availability of three different forms of mathematical knowledge

(primary and middle school content plus a more specialised form) together with

indicators of pedagogical choices provided an excellent opportunity to analyse the

effects of teachers’ capacities on student learning. We did this for the overall student

mathematics test score as well as by individual content areas (numbers, geometry,

etc.), which allowed for a still more precise accounting of how specific teacher –

and teaching – characteristics help determine learning outcomes.

After presenting a conceptual framework for linking the work of the teacher with

student learning in mathematics together with an overview of the different forms of

teacher knowledge, this paper introduces the Guatemalan context and the analytical

framework including the sample, data and methods. This is followed by the results

and some conclusions.

1 Seatwork refers to assignments pupils are asked to carry out individually while sitting at their desks.2 Recitation refers to question and answer activities individually and in ‘‘chorus’’ (i.e. the whole class

answers in unison).

176 J. H. Marshall, M. A. Sorto

123

Conceptual framework

Figure 1 provides a general overview of the factors that determine student learning,

adapted in this study to the teaching of mathematics in a developing country. Family

background figures prominently as a direct predictor of learning. But based on the

evidence from quantitative analyses of student test scores (Wayne and Youngs

2003; Fuller and Clarke 1994), the same is not true for teacher background

characteristics like experience, education and training, which are positioned along

the outer edges of Fig. 1. The challenge for researchers – in both conceptual and

empirical terms – is to fill in the chain (shown as arrows in Fig. 1) linking these

background measures with student learning. This in turn requires opening up the

teaching and learning ‘‘black box’’ in order to understand more about the

antecedents of effective teaching.

Teacher capacity refers to specific domains of knowledge that are critical for

good teaching. These domains of knowledge have been studied (both in theory and

empirically) under different conceptualisations and terminology. Shulman (1986)

first introduced a classification of teacher knowledge into three categories: (1)

subject matter content knowledge; (2) pedagogical content knowledge (PCK)

combining elements of content and teaching knowledge; and (3) curricular

knowledge. Shulman’s PCK concept has since been expanded and used in many

contexts, although often with different meanings (e.g. Grossman 1990 in literature;

Wilson and Wineburg 1988 in social sciences; Watson 2001 in statistics education).

Heather Hill, Deborah Ball and Stephen Schilling (2008a) addressed this lack of

clarity by introducing an updated terminology for mathematics that expands on the

PCK domain while also defining what it is and how it relates to student outcomes

(see Ball et al. 2005; Hill et al. 2004). They call this mathematical knowledge for

teaching, or MKT. MKT has two main categories: (1) subject matter knowledge,

including common content knowledge (or CCK), specialised content knowledge

(SCK), and ‘‘knowledge at the mathematical horizon’’; and (2) pedagogical content

knowledge with a more explicit classification to include knowledge of content and

students (KCS), knowledge of content and teaching (KCT), and knowledge of

curriculum. There is clearly some overlap with Shulman’s original PCK formula-

tion, but with some important differences. First, the subject matter knowledge in

Shulman’s conceptualisation is a subset of the subject matter knowledge for MKT

and, according to Hill et al. (2008a), it corresponds to the CCK domain only.

Second, Shulman did not consider the knowledge of curriculum as part of the PCK

element, but rather as a separate domain of knowledge.

The framework and terminology introduced by Hill et al. (2008a) underlie a lot of

recent work in the United States. Theirs is certainly not the last word on this

concept, as noted by some recent extensions in an international study of

mathematics (Tatto et al. 2008). Nevertheless, the MKT ‘‘umbrella’’ covers a lot

of terrain for understanding teacher capacity, and it is the conceptualisation that we

incorporated in this study.

Returning to Fig. 1, teacher capacity begins with common content knowledge

(CCK) and the necessity that teachers are familiar with the subject matter they are

responsible for (Boero et al. 1996). These minimum skills should not be assumed,

The effects of teacher mathematics knowledge and pedagogy 177

123

especially in developing countries. For example, Ralph W. Harbison and Eric A.

Hanushek (1992) encountered primary school teachers in rural northeast Brazil who

actually scored lower on mathematics tests than their students. However, with the

steady expansion of teacher certification and training, this kind of result, even in the

poorest countries, appears to be becoming less likely (see Passos et al. 2005).3 CCK

also includes the teacher’s content knowledge of higher levels of mathematics.4

Teachers with a profound understanding of mathematics – meaning they are

comfortable beyond the level they are teaching – are better equipped to tackle the

day-to-day work of mathematics instruction. For instance, higher level knowledge is

useful for detecting and anticipating student mistakes and misconceptions, although

content knowledge alone is not likely to help them provide effective feedback.

Specialised content knowledge (SCK) goes beyond knowledge of content and is

useful for specific teaching moments. Examples include the explanations that

teachers use to develop a deep understanding of concepts that are part of the

curriculum, the ways in which they make connections horizontally with other

elements of mathematics at that level, and the questions they set students. However,

in practice specialised content knowledge is more often measured on paper based on

teacher responses to questions than based on observations; this too is changing as

researchers develop more advanced protocols for observing the work of teachers

(Hill et al. 2008a; Sorto and Sapire 2011). As a result, specialised content

knowledge (SCK) is an additional element of teacher capacity in Fig. 1, albeit a

STUDENT LEARNING

PEDAGOGICAL CHOICES

TEACHER CAPACITY

CAPACITY UTILISATION

SYSTEMIC FACTORS

TEACHER BACKGROUND

RESOURCES

PEER EFFECTS

FAMILY BACKGROUND

Fig. 1 A model of mathematics learning

3 A lot of this work has been generated by the Southern and Eastern Africa Consortium for MonitoringEducational Quality (SACMEQ). This project has collected extensive data on student and teacher content

knowledge in a diverse group of African countries. Their results consistently find that teachers score

much higher than their students in elementary school mathematics. For data and studies see

www.sacmeq.org.4 In their report entitled The Mathematical Education of Teachers, the Conference Board of the

Mathematical Sciences (2001), the American Mathematical Society and the Mathematical Association of

America recommend ‘‘a thorough mastery of the mathematics in several grades beyond that which they

expect to teach, as well as of the mathematics in earlier grades’’(p. 7).

178 J. H. Marshall, M. A. Sorto

123

potentially more powerful one compared with common content knowledge. This

distinction between hypothetical and applied knowledge may seem inconsistent

with a concept that is so grounded in application. But teacher trainers and

researchers have long been puzzled by the dynamic of the ‘‘expert student’’ who

becomes a ‘‘novice teacher’’ (Shulman 1986; Borko et al. 1992; Eisenhart et al.

1993). This in turn raises the possibility that teachers with apparently high levels of

mathematics knowledge may not be effective in actually delivering that curriculum

in the classroom.

One explanation for ‘‘underperforming’’ teachers is that high levels of content

and specialised knowledge may be of little help when the teacher lacks the general

pedagogical skills required to create an environment where learning can take place

(meaning there are other forms of knowledge that matter as well). But this

discussion so far largely assumes that pedagogical actions are determined solely by

capacity. In reality the choices that teachers make in the classroom are affected by

what they can do (i.e. capacity) as well as their motivation to fully apply these skills.

This is captured in Fig. 1 by the capacity utilisation element, which includes

systemic and environmental features that affect teacher behaviour.

A substantial body of evidence demonstrates the potential for school supervision

and support regimes to impact the work of teachers in the classroom (Rizvi 2008;

Anderson 2008; Sargent and Hannum 2009). Nevertheless, many teachers in the

developing world work in very isolated situations, with little support or account-

ability. This may help explain in part the high rates of school closure and teacher

absences observed in places like rural Guatemala (Marshall 2009). But based on

recent research in classrooms in Honduras (UMCE 2003), Panama (Sorto et al.

2009), Brazil (Carnoy et al. 2007) and South Africa (Sorto and Sapire, 2011) there is

good reason to be concerned about a pedagogical manifestation of this problem as

well. Teachers in these studies were frequently observed relying on students

copying activities from the blackboard or workbooks, with extended periods of

individual seatwork devoted to procedural exercises with little instruction.

Recitation activities were often limited to very simple, yes/no questions, and

students were rarely asked to justify their answers or make a connection with other

elements of mathematics. Student mistakes were not always corrected, and were

often marked as incorrect and then the student was told to simply fix the problem.

Manipulatives5 and visual models were rarely used.

There are clearly teacher capacity limitations in this situation, meaning teachers

may lack basic content knowledge or the necessary pedagogical content knowledge

to articulate effective explanations when students struggle. In many cases their

pedagogical choices may simply reflect the kind of training they received, or the

way they were taught when they were in school. But there is also the possibility that

these teachers are not fully applying themselves in their work, which in turn opens

the door to a range of school environmental and support influences that need to be

considered together with teacher capacity as explanations of low quality teaching.

This discussion of the teacher’s impact on student achievement in mathematics

highlights three research areas where much work remains. The first requires

5 Manipulatives are hands-on models, e.g. geometric shapes, designed for learners of mathematics.

The effects of teacher mathematics knowledge and pedagogy 179

123

information on how pedagogical choices affect student achievement. Ideally these

data would capture the critical components of effective maths instruction (Hill et al.

2007; Seidel and Shavelson 2007). But even basic indicators of pedagogy – when

based on actual observations – are likely to be useful for understanding why some

teachers are more effective than others. Second, how are these choices themselves

influenced by teacher capacity, including their knowledge of mathematics, and the

work environment? The studies by Hill et al. (2005) and Baumert et al. (2010) make

a convincing case that the teacher’s mathematics knowledge matters; but there is

still the need to identify the pedagogical mechanisms by which this knowledge is

transferred to specific teaching moments. Finally, how are different forms of teacher

knowledge (CCK, SCK, etc.) related to student learning? Despite the growing

interest in this topic the research base is still pretty limited, which in turn highlights

the need for more theory testing and applications of these ideas around the world.

Analytical framework

Guatemalan context

Guatemala has a population of 13 million people, and is located in the Central

American region bordering Mexico, El Salvador, Honduras and Belize. It is one of

the poorest countries of Latin America. 25 per cent of the population lives on less

than the equivalent of two USD per day, and at the time of this data collection

(2002) approximately 70 per cent of the rural population lived below the poverty

line defined by the World Bank (UNDP 2009; World Bank 2002). Parents often

have less than three years of completed education, and many mothers are illiterate.

Child labour indices are also among the highest in the region (Marshall 2003).

Guatemala has a rich cultural heritage, and is one of the few places in Latin

America where large numbers of indigenous peoples have maintained their

language and dress. But these communities have also received little from the central

government in terms of public services (McEwan and Trowbridge 2007), in part

because of their concentration in rural areas. This situation is slowly changing,

stimulated by the official recognition of bilingual education (in 1996) and a

redoubling of efforts to improve rural education. At the time of this survey the

national net primary enrolment rate stood at 88 per cent, although this is lower in

rural areas (UIS 2007).

Given the widespread poverty in these communities, a policy and research

emphasis on the work of the teacher is well justified. A recent sector study (PREAL

2008) highlights both the accomplishments in coverage and the challenges that

remain to guarantee a quality education for all students. The structural constraints

are considerable, as per pupil spending at the primary level (in 2006) was at about

430 US dollars per year (PREAL 2008, Table 20).6 Teachers are graduates of

normal schools that provide three years of high school-level instruction, and in rural

6 Primary schooling in Guatemala has six grades, with the official age of entry being seven.

180 J. H. Marshall, M. A. Sorto

123

areas especially few primary school teachers have completed post-secondary

education.

Sample and data collection

The data collection began at the end of the 2001 school year (August–September)

with a set of Spanish and mathematics exams in a nationally representative sample

of rural grade three classrooms. This work was completed by the ProgramaNacional de Evaluacion del Rendimiento Escolar (PRONERE) evaluation project

(PRONERE 2001), which selected up to thirty grade three students at random in

each school. The follow-up data collection took place throughout the 2002 school

year in a sub-sample of 55 schools drawn from the PRONERE sample in three

administrative states. The states were selected in order to cover the three main

community types in rural Guatemala: largely indigenous Alta Verapaz in the

northern highlands, largely ladino (non-indigenous) Escuintla along the southern

coast, and Chimaltenango in the central highlands, where both indigenous and

ladino populations reside. In each state all of the PRONERE sample schools from

2001 were re-visited. The averages for 2001 test scores and parental education in

these 55 schools are very similar to the national PRONERE averages. This of course

does not mean that the sample is nationally representative, or that the results are

generalisable to all of rural Guatemala. But it is important to note that the schools

were not drawn solely from a single region in what is linguistically and ethnically a

very diverse country.

Two teams of data collection personnel were hired and supervised by the first

author to complete the fieldwork. Each school was visited at least twice throughout

the 2002 school year. The first visit, which took place between April and June, was

used to observe teacher work in the classroom, update student lists from the

previous year’s PRONERE test application, and distribute the first of two

questionnaires to teachers. The schools were then revisited during a two-week

period in late July/early August for conducting the same PRONERE exams with the

cohort of students who had originally participated in the 2001 test.7 Additional data

on student and family background were obtained through a 15–20-minute interview

with each student after the test application, and teachers were asked to complete the

second questionnaire while their students worked on the exams. The late July/early

August period was chosen in order to maximise the number of students who were

available for testing. Guatemalan schools officially begin their school year in

January and end it in October, but in rural areas the actual calendar is often shorter,

and students can begin to leave in August or September.

In most cases the roughly 900 tested students from 2001 were in grade four in

2002, with a smaller group (about 10 per cent) of repeaters who remained in grade

three. Data collection personnel were instructed to distribute the teacher question-

naires to all teachers with students in the testing cohort, regardless of grade (or

section). This work was facilitated by several factors. First, in 12 schools the same

7 There were two slightly different forms (see the Variables section in this article). Students were given

the opposite form (A or B) the second time round.

The effects of teacher mathematics knowledge and pedagogy 181

123

teacher was responsible for grades three and four, while in another 15 schools there

were no (tested) grade three repeaters in 2002. Also, only six of the 55 schools had

more than one grade four section. About 90 teachers completed the background

questionnaires measuring things like experience and maths content knowledge.

However, in most schools the classroom observations were conducted in only one

grade four classroom per school. This does result in missing data for some students,

a detail which we will return to below.

Variables

Table 1 lists the variable names, definitions, means and standard deviations (when

appropriate). Controls for student/family background include student gender, age,

frequency of school attendance, ethnicity, parental education and the student’s

incoming (2001) mathematics score. These data confirm the widespread poverty in

rural Guatemala, as evidenced by parental education levels of roughly two years.

There are some additional controls for school and community characteristics, as

well as the four main groups of variables used in this study.

Student achievement measures (listed under Achievement-dependent variables in

Table 1) include whole test summaries (from 2001 and 2002) and seven specific

content area scores for the 2002 test. The test questions were created by PRONERE

subject specialists in multiple choice format. The two forms (A and B) included

different questions, and were intended to be symmetrical with each item correspond-

ing to a nearly identical question on the opposite form.8 However, for a handful of

items the difficulty levels (i.e. per cent who answered correctly) are significantly

different between the two forms. These differences do not turn up in comparisons of

the overall averages by form, and the PRONERE tests have very good properties in

nationally representative samples (PRONERE 2001). Nevertheless, to further

strengthen the test form equivalence we used a two-parameter item response theory

(IRT) model to exclude several items with poor fit (Cartwright 2010; Crocker and

Algina 1986). The remaining item characteristics were then used to generate adjusted

student percentage correct scores for the overall test (in 2002) and by content area.9

The student scores in Table 1 are presented as percentages between 0 and 100.

The most important result is that there is clear improvement taking place on the

mathematics exam between the end of the 2001 school year (when the PRONERE

tests were applied) and the end of the 2002 school year (when the sub-sample was

revisited). Average achievement increased from 43.8 per cent to 57.5 per cent

during this period. The results by content areas suggest that students were most

comfortable with addition-subtraction and units, and struggled with fractions. But

8 For example, on Form A the second addition task is 352 ? 234 ? 601, whereas on Form B the

corresponding task is 351 ? 241 ? 602.9 This was implemented using the Item and Test Analysis (IATA) programme (version 3.0); see

Cartwright (2010) for details about the software and model. Item pairs with statistically indistinguishable

difficulty levels by form were treated as a single anchor item answered by all students. The remaining

items were included as separate questions answered by half of the students. Items with discrimination

values below 0.30 were discarded for the final analysis. The IRT-generated percentage correct measures

are very similar to the raw averages.

182 J. H. Marshall, M. A. Sorto

123

Table 1 Variable definitions, means and standard deviations

Variable Definition Mean Standard

Deviation

Achievement-dependent variables

Whole Test 2002 Percentage correct for all 53 questions on student

exam applied in 2002 (Alpha = 0.91)

57.5 15.9

Addition–subtraction Percentage correct for 8 addition–subtraction items

(Alpha = 0.59)

79.6 18.6

Multiplication–division Percentage correct for 10 division items

(Alpha = 0.83)

56.4 35.4

Geometry Percentage correct for 7 geometry items

(Alpha = 0.51)

48.8 25.1

Fractions Percentage correct for 3 fractions items

(Alpha = 0.62)

38.7 27.6

Units Percentage correct for 5 units items (Alpha = 0.65) 71.3 29.2

Meaning of operations Percentage correct for 7 understanding operations

items (Alpha = 0.77)

50.8 28.6

Problem solving Percentage correct for 20 problem solving items

(Alpha = 0.76)

46.8 20.5

Student family background

Whole Test 2001 Percentage correct on ‘‘incoming’’ 2001 maths test 43.8 15.3

Student age Student’s age in years 11.6 1.5

Female 1 = Student is female, 0 = male 0.48 –

Indigenous 1 = Student reports speaking a Mayan language;

0 = no

0.67 –

Parental education Average years of parental education (of student) 2.4 2.2

Grade 4 1 = Student in grade 4 in 2002; 0 = student in

grade 3

0.92 –

Teacher–school characteristics

Common content knowledge

(CCK primary)

Teacher percentage correct for 18 items taken from

primary school mathematics curriculum

91.0 8.1

Common content knowledge

(CCK middle school)

Teacher percentage correct for 16 items taken from

middle school mathematics curriculum

53.2 18.3

Teacher specialised content

knowledge (SCK)

Teacher percentage correct out of 8 total points on

three activities

64.8 21.1

SCK activity 1 Raw Score (0–2 points) on item 1 1.5 0.8

SCK activity 2 Raw Score (0–2 points) on item 2 1.4 0.8

SCK activity 3 Raw Score (0–3 points) on item 3 2.4 0.8

Mathematics knowledge for

teaching (MKT)

Teacher percentage correct on all common and

specialised content knowledge items

71.0 11.2

Teacher experience Number of years of overall experience 8.8 7.2

Teaching segments Based on class observations, refers to percentage of

total class time devoted to each segment

Copying/student

seatwork?

Time spent (%) copying and in procedural exercises 40.0 22.2

The effects of teacher mathematics knowledge and pedagogy 183

123

some caution is required in comparing across content areas, and for the areas of

units and fractions the IRT programme was not able to create a content area-specific

score (the raw percentage correct is used instead). We were also unable to obtain the

original item answers for each student from 2001, so it is not possible to extend the

IRT analysis to the incoming test or calculate area-specific gain scores. Table 1 also

includes the number of items in each content area together with Cronbach’s Alpha

measure of test (or ‘‘sub-test’’) reliability.

The teachers’ common content knowledge (CCK) was measured with 18 items

drawn from primary level mathematics, augmented by 16 items drawn from the

grade seven curriculum. Because the teachers and students all answered five anchor

items from the student test, the IRT analysis was extended to obtain a comparable

score (with students) of teacher content knowledge. The IRT-scaled percentage

correct is about 67 per cent, or roughly 0.80 standard deviations higher than the

average student score. This is a rare opportunity to compare teacher and student

content knowledge in a developing country setting, and the result suggests at least

some minimal mathematics preparation for these teachers. Overall they averaged

roughly 90 per cent correct for the primary school level items, compared with about

65 per cent for students on these same questions. However, for the middle school-

Table 1 continued

Variable Definition Mean Standard

Deviation

Teacher check Time spent (%) teacher checking student work

while students work

7.2 7.0

Question–answer (Ind.) Time spent (%) in Individual question and answer 6.8 6.6

Question–answer (Group) Time spent (%) in group (‘‘chorus’’) question and

answer

6.2 7.2

Student at board Time spent (%) with student(s) at blackboard

working

6.4 8.1

Group work Time spent (%) with students working in groups 2.8 5.8

Teacher-led instruction Time spent (%) in teacher instruction activities 27.2 15.4

Transitions Time spent (%) in transitions or interruptions 4.2 3.0

Class size Number of students in classroom 32.3 7.4

Average education School average parental education 2.2 1.1

State controls

Alta Verapaz? 1 = School located in Alta Verapaz; 0 = No 0.56 –

Chimaltenango 1 = School located in Chimaltenango; 0 = No 0.29 –

Escuintla 1 = School located in Escuintla; 0 = No 0.15 –

Source Author data, 2003

Notes Raw averages for student achievement dependent variables (in parentheses) refer to IRT scaled

percentage correct (0–100); SD is for standard deviation; Alpha is Cronbach’s Alpha for reliability of

items for each content area. For Maths Test 2001 the only result available is the percentage correct on the

overall test. Time segments are based on observation of a single mathematics lesson, and refer to per cent

of class time spent in each activity. ? refers to excluded category in statistical comparisons

184 J. H. Marshall, M. A. Sorto

123

level content knowledge questions these primary school teachers clearly struggled:

the overall average correct is only about 50 per cent (see Table 1).10

The teachers’ specialised content knowledge (SCK) in mathematics was assessed

using three items created by the second author. These open-ended questions asked

teachers to diagnose student errors and create a word problem; the items and grading

schematics are presented in more detail elsewhere.11 The questions are intended to

capture the teacher’s applied knowledge by embedding content knowledge elements

within specific teaching situations. Our SCK measure lacks the kind of ‘‘breadth and

depth’’ that comes from applying IRT analysis to a large number of items over a

wide range of teaching areas (see Hill et al. 2005; Baumert et al. 2010). But the

availability of this kind of specialised measure together with common content

knowledge provides a useful opportunity to compare the relative importance of

different forms of knowledge. The overall average is about 65 per cent correct (or

roughly 5 points out of the total 8). Table 1 also presents the results for each of the

three individual activities that make up the overall SCK score.

The three separate knowledge measures for primary school content, middle

school content and specialised content knowledge were also combined to create a

single indicator of mathematics knowledge for teaching (MKT) (see Table 1). It

should be noted that our MKT measure covers only content knowledge in common

and specialised form, which is similar to the strategy employed by Heather Hill,

Brian Rowan and Deborah Ball (2005) in their study of student learning in United

States elementary schools. This means we cannot touch on the full range of the

MKT concept as defined by Hill et al. (2008a). Nevertheless, the overall measure

goes beyond most previous studies, and we also retain the flexibility to consider the

effects of different forms of knowledge, as in Baumert et al. (2010).

Data on classroom processes were collected by observing a single grade four

mathematics lesson during the middle of the school year. The main instrument

applied was a time segment summary, based on a simple ‘‘time on task’’ framework

(Carroll 1963). Data collection personnel observed an entire mathematics lesson and

indicated the predominant segment during each 15-second period. The enumerators

were trained by the first author at the beginning of the school year using videotapes

of classes from other countries, as well as ‘‘live’’ practice sessions in schools to

compare notes. The actual observations were completed individually in real time,

and the classes were not filmed. Each lesson can be summarised as a series of

percentages devoted to individual activities, summing to 1 (or 100 per cent).

The summaries in Table 1 show that the predominant activity is student seatwork

(40 per cent of all time), divided into copying instructions and working on

procedural exercises individually. A version of this segment (‘‘Teacher Checking’’)

10 Of the two measures of common content knowledge (CCK) the fairly high mean (together with low

variance) for the primary level limits its utility for classifying teachers, especially if those with lower

scores were missing questions because of carelessness. We investigated this further using IRT and factor

analysis; the results identified a handful of items with poor fit (mainly at the primary level). But the power

of these statistical extensions is limited somewhat by sample size, so in the empirical work below the

teacher mathematics knowledge indicators refer to raw percentages (or total points), and are not adjusted

using IRT.11 Grading schematics are available upon request from the authors.

The effects of teacher mathematics knowledge and pedagogy 185

123

refers to students engaged in seatwork while the teacher is circulating to check their

work. Recitation activities include question and answer activities individually and in

‘‘chorus’’ (i.e. the whole class answers in unison), as well as students working at the

blackboard. Teacher-led instruction takes up about 27 per cent of the average class,

which includes lecturing, explaining, and demonstrating an example on the board.

Finally, transitions refer to ‘‘dead time’’ in between activities, interruptions and

episodes of disciplining a student (which were rarely observed).

These kinds of classroom snapshots are far from ideal for capturing the

mathematical quality of the lesson. Instead they tell us more about the general

strategies employed by teachers, and the extent to which they are directly involved

in the learning activities in the classroom. We have no way of verifying that each

observed teaching outcome is a valid representation of how these teachers teach

every day. Nevertheless, we are sceptical that things are much different on other

days, for three reasons. First, the overall flavour of these results is consistent with

our observations of classrooms elsewhere in rural Guatemala, and throughout the

Central American region (UMCE 2003). Second, it seems unlikely that teachers are

able to significantly change their teaching style given the fairly limited preparation

they receive, and in most cases permission to observe the class was obtained on the

same day the observation took place. And finally, the students themselves play a

part in these outcomes, and their behaviour is not likely to be easily modified by

teachers in response to having visitors.

Methodology

The student test score data are hierarchical since students are grouped together (or

nested) within individual classrooms, mainly in the grade four classroom in 2002.

Multilevel models are popular in education because they explicitly account for

this kind of nesting by estimating parameters that correspond to different levels

(Raudenbush and Bryk 1986). In the present case there are two levels of observation,

corresponding to students (level 1) and teachers (level 2); a three-level model with

teachers nested within schools is not necessary because these rural schools on

average are very small, and have one section where most of the students are found.

We use the Hierarchical Linear and Nonlinear Modeling (HLM) programme (Bryk

and Raudenbush 1992) to estimate achievement models of the form:

yij ¼ p0j þ p0X Xð Þiþeij ð1Þ

p0j ¼ b00 þ b0MK MKTð Þjþb0TS TSð Þjþb0T Tð Þjþb0S Sð Þjþr0j ð2Þ

Maths achievement y for student i is measured as an overall score as well as for

seven specific content areas (see Table 1). The level-one predictors include a

teacher- (or classroom-) specific intercept (p0j), a vector of individual student and

family background characteristics (X), and an error term.12 Among these background

12 For the final analyses, all of the level 1 predictors (except the intercept) are fixed, meaning they do not

have a random component. In some preliminary analyses significant random effects were encountered,

but these estimations were not very robust and varied considerably by content area. So the more

conservative estimations are presented here.

186 J. H. Marshall, M. A. Sorto

123

variables is the student’s score on the 2001 PRONERE mathematics exam, which is

used for the overall test score model as well as for the content area-specific

estimations. As noted before, it was not possible to apply IRT scaling to the 2001

results, so gains are captured indirectly using the level of mathematics knowledge at

the end of 2002 while controlling the overall level score from the end of the 2001

school year.

Equation 2 generates the various point estimates b for the level-2 predictors of

the (adjusted) classroom average achievement (p0j). The teacher mathematics

knowledge for teaching (MKT) vector includes common content knowledge (CCK),

specialised knowledge (SCK) and an overall indicator (see Table 1); each is used in

different specifications that are described in more detail in the Results section

below. The only other teacher background variable (corresponding to T in Eq. 2)

that is included is for experience. In preliminary analyses the teacher’s education

level was used as well, but there is some collinearity between education and

mathematics knowledge (this is discussed later). This is a somewhat parsimonious

specification for the teacher, but the availability of mathematics knowledge

measures and observational data on teaching means we are not dependent on the

kinds of teacher background indicators incorporated in most existing studies.

The teaching segment (TS) summaries are taken from the classroom observa-

tions. Since the segments add up to 100 per cent they are interpreted in much the

same way as a group of categorical variables for race or school type. The coefficient

for each included segment (Question–Answer, etc.) is interpreted in relation to the

excluded category (Student Seatwork). These are not 0–1 measures corresponding

to the predominant overall strategy employed in the observed class. They are instead

linear measures corresponding to the percentage of the lesson that was spent in each

instructional activity. So each point estimate is interpreted as the change in maths

achievement given a standard deviation increase in this activity relative to seatwork.

Finally, the school level controls (S) include a measure for class size, average

parental education and state dummies (fixed effects); these help improve the causal

properties of the model overall by controlling for things that may be related to the

distribution of teachers across these communities.

Before estimating the final models we had to deal with missing data. Only about

80 of the roughly 900 students with 2001 and 2002 test score data are missing

individual or teacher background information like parental education, attendance

frequency and teacher experience. But the questionnaires measuring teacher content

and specialised teaching knowledge have more missing data. The available student

sample is reduced by about 150 cases (or 17 per cent of the total) after matching

students with their teacher’s mathematics knowledge; this corresponds to losing

roughly 20 of the original 90 teachers (or 22 per cent of the teachers). Also, there is

another group of roughly 20 teachers for whom we have teacher mathematics data

but incomplete observation data; this corresponds to another 150 students.

We considered three options for handling missing data. Our preferred approach

matched students with teachers for the teacher mathematics knowledge data, but

computed school averages for the classroom observations to fill in for missing

observation data. This resulted in roughly 70 teachers and 700 students, or roughly

85 per cent of the original sample for whom complete student data were available.

The effects of teacher mathematics knowledge and pedagogy 187

123

We also weighted the data for non-response, although the response function shows

no significant student or school background predictors of having missing data.

Results

Table 2 presents the results for the HLM analyses of student achievement. The

coefficients for continuous variables measure one standard deviation changes.

Additional variables that are not presented include controls for student attendance,

parental education and state-fixed effects (see notes underneath Table 2). As

expected, the overall score on the 2001 maths test is a very significant predictor of

the overall and content area-specific mathematics outcomes in 2002, with

standardised effects of upwards of 0.66 standard deviations. Girls score significantly

lower than boys on the overall test, although the differences are concentrated in only

three areas: problem solving, measurement, and meaning of operations. Grade four

students score moderately higher than their grade three counterparts. Part of this

advantage is curriculum-related, since concepts like division are usually introduced

in grade four. One surprise is that indigenous student scores are not significantly

different from Spanish-only speakers (ladinos) when controlling for background,

learning context and incoming scores. This does not mean that their achievement

levels are the same – indigenous students score lower – and it should also be noted

that the sample does not include urban areas where most ladino children reside.

For the overall test score, the teacher mathematics knowledge measures are

introduced one by one (Table 2, columns 1–4). The results show that both of the

level-specific common content knowledge (CCK) measures (primary and middle),

as well as the specialised content knowledge measure (SCK), are significantly

associated with student achievement. The standardised effect sizes are only about

0.05–0.08 standard deviations, although with the control for 2001 achievement the

parameters are measuring a form of learning gain.13 The most significant predictor

is for the overall teacher mathematics knowledge for teaching (MKT) measure that

combines the common and specialised knowledge forms. Interestingly, the effect

size for this variable is almost identical to what Hill et al. (2005) encountered in

their study using a similar measure in United States primary schools.

For the remaining content-area specific estimations in Table 2 we use this overall

measure of the teacher’s mathematics knowledge for teaching (MKT). The point

estimates are positive for each of the seven content areas, although only significant

in problem solving (moderately) and fractions. Overall, these results provide some

empirical support for a widely held, if infrequently tested, belief in mathematics

education: effective teachers have different kinds of mathematical knowledge.

However, we can only speculate about the underlying teaching mechanisms that

link mathematics knowledge with higher student test scores, in part because of the

lack of correlation between the teaching segment data and teacher background (we

13 In other estimations (not presented) we computed gain scores subtracting the 2001 raw percentage

correct from the IRT-scaled percentage correct in 2002. The results are nearly identical, but do show

moderately larger effect sizes for the estimations using teacher knowledge (columns 1–4 in Table 2).

188 J. H. Marshall, M. A. Sorto

123

Ta

ble

2H

iera

rch

ical

Lin

ear

Mo

del

(HL

M)

esti

mat

eso

fco

var

iate

so

fst

ud

ent

achie

vem

ent,

wh

ole

test

and

spec

ific

mat

hem

atic

sco

nte

nt

area

s

Indep

enden

tvar

iable

Whole

test

Pro

ble

m

solv

ing

Conce

pts

,M

eanin

gof

oper

atio

ns

Alg

ori

thm

s

(1)

(2)

(3)

(4)

Unit

sM

eanin

gof

ops.

Fra

ctio

ns

Geo

met

ryA

dd–

subtr

act

Mult

iply

div

ide

Lev

el1

pre

dic

tors

Whole

test

2001

0.6

5***

(18.8

4)

0.6

6***

(18.9

6)

0.6

6***

(18.9

9)

0.6

4***

(18.9

1)

0.5

4***

(15.3

0)

0.3

4***

(11.2

3)

0.4

8***

(15.4

9)

0.2

7***

(7.1

6)

0.2

8***

(6.6

5)

0.3

8***

(6.8

1)

0.4

3***

(12.7

1)

Fem

ale

-0.1

5***

(-3.3

1)

-0.1

5***

(-3.2

9)

-0.1

4***

(-3.2

3)

-0.1

5***

(-3.2

8)

-0.1

0*

(-1.9

1)

-0.1

8***

(-2.5

0)

-0.2

2***

(-3.1

1)

0.0

4

(0.5

8)

-0.0

4

(-0.3

9)

-0.0

1

(-0.0

2)

-0.0

8

(-1.1

9)

Indig

enous

0.0

3

(0.2

4)

0.0

2

(0.1

6)

0.0

3

(0.2

2)

0.0

2

(0.1

9)

0.0

9

(0.5

9)

0.1

8

(-1.4

7)

-0.1

6

(-1.5

9)

0.0

1

(0.0

3)

-0.0

9

(-0.6

6)

0.2

5

(1.4

0)

0.1

8

(0.9

6)

Gra

de

40.1

9*

(1.6

5)

0.1

8

(1.5

3)

0.1

7

(1.5

0)

0.1

8

(1.5

3)

-0.0

7

(-0.4

1)

0.2

9

(1.6

0)

0.2

1

(1.3

7)

0.1

1

(0.6

8)

0.3

8***

(3.1

0)

0.0

1

(0.0

3)

0.2

8

(1.5

9)

Lev

el2

pre

dic

tors

Tea

cher

mat

hem

atic

sknow

ledge

CC

K(P

rim

ary)

0.0

8**

(2.1

3)

CC

K(M

iddle

)–

0.0

6*

(1.9

0)

––

––

––

––

Spec

iali

sed

conte

nt

know

ledge

(SC

K)

––

0.0

6*

(1.7

9)

––

––

––

––

Over

all

mat

hs

know

ledge

(MK

T)

––

–0.0

7**

(2.2

4)

0.0

7*

(1.7

8)

0.0

5

(1.4

7)

0.0

4

(1.3

6)

0.1

0***

(3.4

6)

0.0

6

(1.3

7)

0.0

2

(0.3

1)

0.0

7

(1.2

6)

Tea

cher

exper

ience

0.0

1

(0.0

9)

0.0

1

(0.2

9)

0.0

3

(0.8

6)

0.0

2

(0.5

0)

-0.0

1

(-0.2

1)

0.0

4

(1.3

9)

-0.0

1

(-0.0

4)

0.0

8**

(2.2

7)

0.0

3

(0.6

9)

-0.0

2

(-0.6

0)

-0.0

2

(-0.4

5)

Tea

chin

gse

gm

ents

a

Tea

cher

chec

kin

g0.1

2***

(2.7

2)

0.1

2***

(2.7

4)

0.0

8*

(1.9

1)

0.1

1***

(2.5

5)

0.1

0*

(1.6

6)

0.0

3

(0.5

9)

0.0

6

(0.9

1)

0.0

5

(1.1

9)

0.0

6

(1.0

4)

0.0

6

(0.8

7)

0.1

8**

(2.3

8)

The effects of teacher mathematics knowledge and pedagogy 189

123

Ta

ble

2co

nti

nu

ed

Indep

enden

tvar

iable

Whole

test

Pro

ble

m

solv

ing

Conce

pts

,M

eanin

gof

oper

atio

ns

Alg

ori

thm

s

(1)

(2)

(3)

(4)

Unit

sM

eanin

gof

ops.

Fra

ctio

ns

Geo

met

ryA

dd–

subtr

act

Mult

iply

div

ide

Ques

tion–an

swer

(indiv

idual

)-

0.0

2

(-0.3

8)

-0.0

2

(-0.4

8)

-0.0

1

(-0.2

8)

-0.1

1

(-0.3

8)

-0.0

8

(-1.3

9)

0.0

6

(1.3

8)

-0.0

6

(-0.9

6)

0.0

5*

(1.7

3)

-0.0

2

(-0.4

7)

0.0

5

(0.8

9)

0.1

0

(1.5

0)

Ques

tion–an

swer

(‘‘C

horu

s’’)

0.0

2

(0.3

7)

0.0

3

(0.5

5)

-0.0

1

(-0.0

3)

0.0

2

(0.3

9)

-0.0

1

(-0.1

8)

-0.0

1

(-0.2

6)

0.0

4

(0.5

7)

0.0

5

(1.6

2)

0.0

6*

(1.7

4)

0.0

6

(1.0

7)

-0.0

3

(-0.2

0)

Stu

den

tat

boar

d0.0

9***

(2.9

4)

0.0

9***

(2.6

6)

0.1

0***

(3.1

9)

0.0

9***

(2.8

5)

0.0

6

(1.3

4)

-0.0

4

(-0.9

5)

0.0

3

(0.9

5)

0.0

2

(0.6

4)

0.0

2

(0.5

3)

0.0

9*

(1.7

0)

0.0

8

(1.5

8)

Gro

up

work

0.1

6***

(6.0

5)

0.1

4***

(5.7

0)

0.1

5***

(6.0

8)

0.1

5***

(5.6

3)

0.1

6***

(5.2

4)

0.1

1***

(5.2

4)

0.0

4

(0.6

8)

0.0

7**

(2.2

3)

0.0

9**

(2.5

4)

0.0

5

(1.0

8)

0.1

4***

(2.7

7)

Tea

cher

-led

inst

ruct

ion

0.1

4***

(3.0

1)

0.1

4***

(2.9

7)

0.1

4***

(2.7

7)

0.1

4***

(3.0

5)

0.1

5**

(2.7

9)

-0.0

4

(-1.1

2)

0.1

1*

(1.7

8)

0.0

9*

(1.8

6)

0.0

3

(0.3

7)

0.0

9

(1.2

5)

0.1

8**

(2.4

1)

Tra

nsi

tions

-0.0

8**

(-2.0

1)

-0.0

7**

(-2.0

1)

-0.0

7**

(-1.9

7)

-0.0

7**

(-1.9

5)

-0.0

6

(-1.0

3)

-0.1

0***

(-3.1

0)

-0.0

5

(-1.5

8)

-0.0

2

(-1.1

6)

0.0

2

(0.2

8)

0.0

1

(0.0

8)

-0.0

9

(-1.3

9)

N(s

tuden

ts,

clas

sroom

s)(6

99,6

5)

(699,6

5)

(699,6

5)

(699,6

5)

(699,6

5)

(699,6

5)

(699,6

5)

(699,6

5)

(699,6

5)

(699,6

5)

(699,6

5)

Expla

ined

var

iance

(per

cent)

Wit

hin

school

53.5

53.4

53.4

53.4

37.1

27.6

21.9

21.1

15.1

14.0

34.1

Bet

wee

nsc

hool

82.3

82.3

82.9

83.2

57.7

86.9

64.5

88.5

60.0

35.5

45.9

Sourc

eA

uth

or

dat

a,2003

Note

sE

stim

ates

wer

eobta

ined

usi

ng

the

HL

Mpro

gra

mm

e(B

ryk

and

Rau

den

bush

1992

).L

evel

-1ch

arac

teri

stic

sar

efi

xed

,w

ith

the

exce

pti

on

of

the

clas

sroom

-lev

elin

terc

ept

whic

his

model

led

usi

ng

the

level

-2pre

dic

tors

.P

aram

eter

sre

fer

toth

ech

ange

(in

stan

dar

ddev

iati

ons)

inea

chdep

enden

tvar

iable

giv

ena

one

stan

dar

ddev

iati

on

chan

ge

inin

dep

enden

tvar

iable

(when

appli

cable

).

Dep

enden

tvar

iable

mea

ns

and

stan

dar

ddev

iati

ons

are

pre

sente

din

Tab

le1

aF

or

teac

hin

gse

gm

ents

the

excl

uded

cate

gory

refe

rsto

studen

tco

pyin

g/s

eatw

ork

.A

ddit

ional

var

iable

sth

atar

enot

incl

uded

inT

able

2ar

e:st

uden

tag

e,st

uden

t’s

aver

age

par

enta

led

uca

tion,

studen

tat

tendan

cefr

equen

cyduri

ng

2002

school

yea

r,cl

ass

size

,sc

hool

aver

age

par

enta

led

uca

tion

and

stat

e-fi

xed

effe

cts.

Ast

eris

ks

refe

rto

two-t

ail

signifi

cance

level

s(*

**

0.0

1,

**

0.0

5,

*0.1

0)

190 J. H. Marshall, M. A. Sorto

123

return to this below). One interesting clue is that the teacher’s overall mathematics

knowledge for teaching is most significant in areas of the student test that have the

highest levels of cognitive demand. For example, even at primary level problem

solving requires students to put together different aspects of mathematics, and

fractions is one of the more difficult concepts in mathematics because effective

teaching requires linking the algorithms with conceptual meaning.

The time segment summaries come next in Table 2. Most of the coefficients are

positive in relation to the excluded category made up of copying/seatwork. This

result largely confirms our expectation that the least effective schools are those

where student copying and individual seatwork appear to predominate. One problem

with this activity, at least as we have observed its use in places like rural Guatemala,

is that it is often inefficient. Seatwork segments can drag on while the teacher waits

for slower children, and even then it is not unusual for some students to be far

behind at the end. This is exacerbated by the fact that seatwork routines in these

classes are often disconnected from the rest of the lesson. For example, they are

infrequently followed up by intensive and effective discussion activities.

One unexpected result is the general lack of significance for the two forms of

question and answer activities. This segment is marked by the teacher going around

the room asking individual or whole class (chorus) questions, usually in simple yes/

no or basic operations (‘‘two plus two equals?’’) format. One explanation is that the

kinds of recitation activities that take place between students and teachers in these

schools involve very limited cognitive content (Vygotsky 1978). Teachers rarely

challenge students to explain their answers, or make connections with other

elements of mathematics.

This take on the question and answer segment results is indirectly supported by

the positive effects on achievement when students spend more time solving

problems at the blackboard, or engaging in group work. Work at the board provides

opportunities for individual tutoring, and it also provides immediate feedback (see

Marshall et al. 2009 for a similar result from Cambodia). When the student ‘‘gets it

right’’ it builds confidence that helps overall performance. Furthermore, this

segment suggests more actively engaged teachers who are interacting with students

in very specific dimensions. Group work also provides students with opportunities

to have direct interaction with others. In the area of problem solving this is likely to

be especially important, since listening and talking to other students makes it easier

to see different ways to solve problems, and also allows for practising multiple

strategies.

Students studying in classrooms marked by more direct instruction also score

significantly higher overall, and in most of the specific content areas. Observing

direct instruction combining explanations with actual activities is likely to be more

productive than just letting students work on problems alone. Also, the teacher’s use

of this particular activity may be based in part on a better preparation of the lesson,

meaning a more detailed lesson plan. In contrast, teachers who rely more on

seatwork may have less teaching content planned for the lesson.

The results for the time segment summaries show that even relatively simple

observational data on teaching actions can provide some useful insight into

questions related to teacher effectiveness. The effect sizes for the various categories

The effects of teacher mathematics knowledge and pedagogy 191

123

are generally in 0.10–0.15 standardised range, which is substantially higher than the

mathematics knowledge measures. Also, the results are robust to the inclusion of

controls for teacher education (not included in Table 2), experience and commu-

nity-level measures of parental education, etc.

Overall, the results in Table 2 provide an unusually detailed accounting of

specific teacher and teaching characteristics that are associated with higher student

maths scores, both overall and by content area. But our interest in the effects of

teacher mathematics knowledge on student achievement warranted a still more

detailed examination of the data. We therefore re-estimated the content area-specific

models in Table 2 by replacing the teachers’ overall mathematics knowledge for

teaching (MTK) scores with each of the individual knowledge measures, including

the three specialised teaching activities that make up the SCK construct. The

condensed results are presented in Table 3. For the teacher’s common content

knowledge (CCK) the results add little to the earlier findings from Table 2. One

exception is for the area of geometry, which is strongly associated with the teacher’s

CCK at the primary school level.

For the overall SCK measure, the coefficients are generally positive but only

significant in the Multiplication-Division algorithm content area. This is a notable

finding, however, because the SCK activities completed by teachers included a

focus on operations for multiplication and division. As a result it is not surprising

that the SCK construct appears to be most closely related to student achievement in

this particular content area. This finding provides some additional justification for

analysing mathematics achievement by content area. It also highlights the need to

continue developing specialised measures of teacher knowledge and attempting to

link these with student achievement outcomes in those same content areas. With

only a handful of teacher SCK questions, our ability in this study to make those

kinds of specific connections is constrained. But as more and more researchers

develop teacher knowledge items – and make them publicly available – future

research should be able to build considerably on these kinds of initial findings.

Discussion and conclusions

In this paper we use data from rural Guatemala to analyse the effects of teachers on

student mathematics achievement. We are able to shed some light on questions that

motivate mathematics policy discussions around the globe, such as the importance

of different forms of teacher mathematical knowledge, and the effects of pedagogy

on student learning. For policymakers (and teachers) in general the results are

encouraging because they suggest specific mechanisms by which effective teachers

can make substantial impacts on student learning, even in extremely poor contexts.

Our results linking higher levels of teacher mathematics knowledge with higher

student test scores provide empirical support for a proposition that makes intuitive

sense, but has been tested in relatively few instances. The content knowledge results

highlight both the imperative of guaranteeing minimum levels of teacher preparation

(Santibanez 2006) as well as the potential for primary school teachers to apply higher

level content skills in their teaching. The significant effects of the teacher’s

192 J. H. Marshall, M. A. Sorto

123

Tab

le3

Hie

rarc

hic

alL

inea

rM

odel

(HL

M)

esti

mat

esof

covar

iate

sof

studen

tac

hie

vem

ent

insp

ecifi

cm

athem

atic

sco

nte

nt

area

s

Ind

epen

den

tv

aria

ble

(add

ed

sep

arat

ely

)

Pro

ble

mso

lvin

gC

once

pts

,M

eanin

gof

oper

atio

ns

Alg

ori

thm

s

Unit

sM

eanin

gof

ops.

Fra

ctio

ns

Geo

met

ryA

dd–su

btr

act

Mult

iply

–div

ide

Tea

cher

com

mon

con

ten

t

(CC

Kp

rim

ary

)

0.0

4

(1.0

1)

0.0

5

(1.3

1)

0.0

6

(1.5

6)

0.1

0*

**

(2.9

1)

0.1

0*

**

(2.6

0)

0.0

3

(0.3

9)

0.0

6

(1.0

7)

Tea

cher

com

mon

con

ten

t

(CC

Km

idd

lesc

ho

ol)

0.0

6

(1.6

3)

0.0

2

(0.6

0)

0.0

7

(0.5

0)

0.0

8*

**

(2.6

3)

0.0

3

(0.6

8)

0.0

2

(0.4

0)

0.0

3

(0.5

8)

Tea

cher

spec

iali

sed

conte

nt

(SC

K)

0.0

7

(1.5

0)

0.0

5

(1.3

6)

0.0

4

(1.0

4)

0.0

5

(1.0

7)

0.0

3

(0.6

7)

-0

.01

(-0

.09)

0.1

1*

*

(2.0

1)

SC

Kac

tivit

ies

SC

Kac

tivit

y1

-0

.02

(-0

.24)

-0

.04

(-0

.70)

-0

.04

(-0

.83)

-0

.05

(-1

.01)

-0

.03

(-0

.70

)

0.1

8*

**

(3.1

3)

0.1

1*

(1.9

1)

SC

Kac

tivit

y2

0.0

4

(0.6

2)

0.0

9*

*

(1.9

8)

0.0

6

(1.2

7)

-0

.04

(-0

.90)

-0

.01

(-0

.27

)

-0

.17

**

*

(-3

.08)

0.1

1*

(1.8

7)

SC

Kac

tivit

y3

0.0

6

(1.4

2)

0.0

1

(0.2

6)

0.0

2

(0.5

8)

0.1

3*

**

(4.6

6)

0.0

7*

*

(2.0

5)

0.0

1

(0.1

6)

-0

.01

(-0

.97)

So

urc

eA

uth

or

dat

a,2

00

3

Not

esE

stim

ates

wer

eobta

ined

usi

ng

the

HL

Mpro

gra

mm

e(B

ryk

and

Rau

den

bush

19

92);

see

mai

nte

xt

and

no

tes

atb

ott

om

of

Tab

le2.

All

coef

fici

ents

refe

rto

stan

dar

dis

edef

fect

so

fa

stan

dar

dd

evia

tio

nch

ang

ein

the

ind

epen

den

tv

aria

ble

.E

ach

of

the

fou

rin

dic

ato

rsfo

rte

ach

erm

ath

emat

ics

kn

ow

led

ge

was

added

separ

atel

yto

the

mai

nm

od

el.

Fo

rex

amp

le,

the

resu

lts

for

Pro

ble

mS

olv

ing

refe

rto

fou

rse

par

ate

esti

mat

ion

su

sin

gid

enti

cal

mod

els

wit

hth

eex

cep

tio

no

fth

esp

ecifi

cin

dic

ato

ro

fte

ach

er

mat

hem

atic

sk

no

wle

dge.

Co

mp

lete

resu

lts

are

avai

lab

leu

po

nre

qu

est.

Ast

eris

ks

refe

rto

two

-tai

lsi

gn

ifica

nce

lev

els

(**

*0

.01

,*

*0

.05,

*0

.10)

The effects of teacher mathematics knowledge and pedagogy 193

123

specialised content knowledge (SCK) on overall student achievement gains, as well

as more specific connections within individual content areas, provide tentative

support for the teaching knowledge concept prized by many mathematics education

researchers.

Does one form of teacher knowledge matter more than the other? Given some

partial overlap between the common content and specialised knowledge domains,

and the specificity of some results by content area, this question is impossible to

answer based on these data (see Baumert et al. 2010). Besides, we are reluctant to

treat the content and specialised knowledge elements as competing modes of

understanding teacher effectiveness, or distinct training approaches. In reality they

both play an important role in preparing teachers for the classroom, a position

echoed in a 2005 article by a group of prominent mathematicians and mathematics

education researchers (Ball et al. 2005).

This study also extends the research base on teacher quality by examining the

effects of actual teaching activities. Using observational data on mathematics

classes we find that student maths scores are lower in schools where more time is

spent copying and solving problems individually. Given the apparent predominance

of this activity in areas such as rural Guatemala this gives cause for concern, since at

least some students are likely to fall behind when the seatwork activities are not

followed up by careful review and recitation. The more productive activities instead

seem to be those that are more certain to engage students in the learning process.

These include direct lecture, sending students to the blackboard, and having them

work in groups.

One of the challenges for studies such as this one is to understand the underlying

dynamics that help explain why some teachers appear to be more effective than

others. Compared with the data linking student test scores with teachers, our

information is less complete on this count, although we do provide more detailed

statistical analyses elsewhere.14 Three results stand out. First, the teachers’

specialised content knowledge in mathematics is significantly correlated with their

common content knowledge (r = 0.40). This confirms the relevance of content

knowledge in understanding specialised knowledge, but it is not the case that one is

a substitute for the other. Second, there is almost no correlation between any of the

teacher mathematics knowledge indicators and the observed pedagogical choices

(time segments). This is an important reminder that teacher actions in the classroom

are not automatically determined by capacity. Finally, there is some evidence that

teachers’ access to university courses is associated with higher levels of common

and specialised content knowledge in mathematics. This supports ongoing efforts in

countries like Guatemala to upgrade teacher training levels.

Future research can build on these results by adapting more comprehensive

observation protocols to measure pedagogical choices in the classroom (Sorto et al.

2009; Hill et al. 2008b; Learning Mathematics for Teaching 2006). Our results

suggest that teachers with higher levels of mathematical knowledge are more

effective in the classroom. But the actual teaching mechanisms – if they do exist –

were not identifiable with our observation instruments. Ideally this kind of

14 On this count; results from more detailed statistical analyses are available upon request.

194 J. H. Marshall, M. A. Sorto

123

information will be linked with student achievement results in different content

areas, which is another improvement in our study design compared with most other

studies of mathematics test scores.

References

Anderson, J. B. (2008). Principals’ role and public primary schools’ effectiveness in four Latin American

cities. The Elementary School Journal, 109(1), 36–60.

Ball, D. L., & Bass, H. (2000). Interweaving content and pedagogy in teaching and learning to teach:

Knowing and using mathematics. In J. Boaler (Ed.), Multiple perspectives on the teaching andlearning of mathematics. Westport, CT: Ablex.

Ball, D. L., Hill, H. C., & Bass, H. (2005). Knowing mathematics for teaching: Who knows mathematics

well enough to teach third grade, and how can we decide? American Educator, 2005 (Fall), 14–22.

Baumert, J., Kunter, M., Blum, W., Brunner, M., Voss, T., Jordan, A., et al. (2010). Teachers’

mathematical knowledge, cognitive activation in the classroom, and student progress. AmericanEducational Research Journal, 47(1), 133–180.

Boero, P., Dapueto, C., & Parenty, L. (1996). Didactics of mathematics and the professional knowledge

of teachers. In A. J. Bishop, M. A. Clements, C. Keitel, J. Patrick, & F. K. S. Leung (Eds.),

International handbook of mathematics education (pp. 1097–1121). Dordrecht: Kluwer.

Borko, H., Eisenhart, M., Brown, C. A., Underhill, R. G., Jones, D., & Agard, P. C. (1992). Learning to

teach hard mathematics: Do novice teachers and their instructors give up too easily? Journal forResearch in Mathematics Education, 23, 194–222.

Bryk, A. S., & Raudenbush, S. (1992). Hierarchical linear models: Applications and data analysismethods. Thousand Oaks, CA: Sage Publications.

Carnoy, M., Gove, A., & Marshall, J. H. (2007). Cuba’s academic advantage: Why students in Cuba dobetter in school. Stanford, CA: Stanford University Press.

Carroll, J. B. (1963). A model of school learning. Teachers College Record, 64, 723–733.

Cartwright, F. (2010). Item and test analysis: A software tutorial and theoretical introduction. Accessed

May 2011 from http://polymetrika.com/IATA/.Conference Board of the Mathematical Sciences. (2001). The mathematical education of teachers.

Providence, RI and Washington, DC: American Mathematical Society and Mathematical

Association of America.

Crocker, L., & Algina, J. (1986). Introduction to classical and modern test theory. Belmont, CA:

Wadsworth Group/Thomson Learning.

Eisenhart, M., Borko, H., Underhill, R. G., Brown, M., Jones, D., & Agard, P. C. (1993). Conceptual

knowledge falls through the cracks: Complexities of learning to teach mathematics for

understanding. Journal for Research in Mathematics Education, 24, 8–40.

Fuller, B., & Clarke, P. (1994). Raising school effects while ignoring culture? Local conditions and the

influence of classroom tools, rules and pedagogy. Review of Educational Research, 64, 119–157.

Grossman, P. L. (1990). The making of a teacher. Teacher knowledge and teacher education. New York:

Teachers College Press.

Harbison, R., & Hanushek, E. A. (1992). Educational performance of the poor: Lessons from ruralnortheast Brazil. New York: Oxford University Press.

Hill, H. C., & Ball, D. L. (2004). Learning mathematics for teaching: Results from California’s

mathematics professional development institutes. Journal for Research in Mathematics Education,35(5), 330–351.

Hill, H. C., Ball, D. L., & Schilling, S. G. (2008a). Unpacking pedagogical content knowledge:

Conceptualizing and measuring teachers’ topic-specific knowledge of students. Journal forResearch in Mathematics Education, 39(4), 372–400.

Hill, H. C., Blunk, M. L., Charalambous, Y. C., Lewis, J. M., Phelps, G. C., Sleep, L., et al. (2008b).

Mathematical knowledge for teaching and the mathematical quality of instruction: An exploratory

study. Cognition and Instruction, 26, 430–511.

Hill, H. C., Rowan, B., & Ball, D. L. (2005). Effects of teachers’ mathematical knowledge for teaching on

student achievement. American Educational Research Journal, 42, 371–406.

The effects of teacher mathematics knowledge and pedagogy 195

123

Hill, H. C., Schilling, S. G., & Ball, D. L. (2004). Developing measures of teachers’ mathematical

knowledge for teaching. Elementary School Journal, 105, 11–30.

Hill, H. C., Sleep, L., Lewis, J. M., & Ball, L. B. (2007). Second handbook of research on mathematics

teaching and learning. In F. K. Lester (Ed.), Assessing teachers’ mathematical knowledge. Whatknowledge matters and what evidence counts? (pp. 111–155). Reston, VA: National Council of

Teachers of Mathematics.

Learning Mathematics for Teaching (LMT). (2006). A Coding rubric for measuring the quality ofmathematics in instruction, Tech. Rep. LMT1.06. University of Michigan, School of Education, Ann

Arbor, MI.

Marshall, J. H. (2003). If you build it will they come? The effects of school quality on primary schoolattendance in rural Guatemala. Doctoral dissertation, Stanford University School of Education.

Marshall, J. H. (2009). School quality and learning gains in rural Guatemala. Economics of EducationReview, 28(2), 207–216.

Marshall, J. H., Chinna, U., Nessay, P., Hok, U. N., Savoeun, V., Tinon, S., et al. (2009). Student

achievement and education policy in a period of rapid expansion: Assessment data evidence from

Cambodia. International Review of Education, 55(4), 393–413.

Marshall, J. H., & White, K. A. (2001). Academic achievement, school attendance and teacher quality inHonduras: An empirical analysis. Tegucigalpa, Honduras: UMCE.

McEwan, P. J., & Trowbridge, M. (2007). The achievement of indigenous students in Guatemalan

primary schools. International Journal of Educational Development, 27, 61–76.

Mullens, J. E., Murnane, R. J., & Willett, J. B. (1996). The contribution of training and subject matter

knowledge to teaching effectiveness: A multilevel analysis of longitudinal evidence from Belize.

Comparative Education Review, 40, 139–157.

Passos, A., Nahara, T., Magaia, F., & Lauchande, C. (2005). The SACMEQ project II in Mozambique: Astudy of the conditions of schooling and the quality of education. Southern and Eastern Africa

Consortium for Monitoring Educational Quality (SACMEQ) and Ministry of Education and Culture,

Mozambique.

PREAL. (2008). Education: A challenge of national urgency. Partnership for Educational Revitalization

in the Americas and Center for National Economic Research (CIEN).

PRONERE. (2001). Informe de resultados del Programa Nacional de Evaluacion del RendimientoEscolar. Guatemala City: PRONERE.

Raudenbush, S., & Bryk, A. S. (1986). A hierarchical model for studying school effects. Sociology ofEducation, 59, 1–17.

Rizvi, M. (2008). The role of school principals in enhancing teacher professionalism: Lessons from

Pakistan. Educational Management Administration & Leadership, 36(1), 85–100.

Rockoff, J. E., Jacob, B. A., Kane, T. J., & Steiger, D. O. (2008). Can you recognize an effective teacherwhen you recruit one? National Bureau of Economic Research Working Paper 14485. NBER,

Cambridge, MA.

Santibanez, L. (2006). Why we should care if teachers get A’s: Teacher test scores and student

achievement in Mexico. Economics of Education Review, 25(5), 510–520.

Sargent, T., & Hannum, E. (2009). Doing more with less: Teacher professional learning communities in

resource-constrained primary schools in rural China. Journal of Teacher Education, 60(3), 258–276.

Seidel, T., & Shavelson, R. J. (2007). Teaching effectiveness research in the past decade: The role of

theory and research design in disentangling meta-analysis results. Review of Educational Research,77(4), 454–499.

Shulman, L. S. (1986). Those who understand: Knowledge growth in teaching. Educational Researcher,15, 4–14.

Sorto, M. A., Marshall, J. H., Luschei, T. F., & Carnoy, M. (2009). Teacher knowledge and teaching in

Panama and Costa Rica: A comparative study. Revista Latinamericana de Investigacion enMatematica Educativa, 12(2), 251–290.

Sorto, M. A., & Sapire, I. (2011). The teaching quality of mathematics lessons in South Africa schools.

Journal of Education, 51, 93–114.

Tatto, M. T., Schwille, J., Senk, S., Ingvarson, L., Peck, R., & Rowley, G. (2008). Teacher education anddevelopment study in mathematics (TEDS-M): Policy, practice, and readiness to teach primary andsecondary mathematics. Conceptual framework. East Lansing, MI: Teacher Education and

Development International Study Center, College of Education, Michigan State University.

UNESCO Institute for Statistics (UIS). (2007). World education indicators 2007. Montreal: UIS.

Accessed 15 June 2007 from www.unesco.org.

196 J. H. Marshall, M. A. Sorto

123

Unidad Externa de Medicion de la Calidad de la Educacion (UMCE). (2003). Factores asociados alrendimiento academico de los alumnos de la Educacion primaria en Honduras 2002. [Factors

associated with primary school academic achievement in Honduras 2002]. Published by UMCE

project. Tegucigalpa, Honduras.

United Nations Development Programme (UNDP). (2009). Human Development Report 2009. New

York: UNDP.

Vygotsky, L. S. (1978). Mind in society: The development of higher psychological processes, Translation.

M. Cole, V. John-Steiner, S. Scribner, & E. Souberman (Eds.). Cambridge, MA: Harvard University

Press.

Watson, J. M. (2001). Profiling teachers’ competence and confidence to teach particular mathematics

topics: The case of chance and data. Journal of Mathematics Teacher Educator, 4(4), 305–337.

Wayne, A. J., & Youngs, P. (2003). Teacher characteristics and student achievement gains: A review.

Review of Educational Research, 73(1), 89–122.

Wilson, S. M., & Wineburg, S. (1988). Peering at history through different lenses: The role of disciplinary

perspectives in teaching history. Teachers College Record, 89, 525–539.

World Bank. (2002). Project appraisal document (Guatemala). Washington, DC: World Bank.

The authors

Jeffery H. Marshall is a Visiting Researcher at the Instituto de Investigacion y Evaluacion Educativas y

Sociales (Institute of Educational and Social Research and Evaluation) at the Universidad Pedagogica

Nacional Francisco Morazan (Francisco Morazan National Teachers University) in Tegucigalpa,

Honduras, and Co-Director of Sapere Development Solutions. Dr Marshall holds a PhD in education from

Stanford University. His research interests are in education policy, programme evaluation, economics of

education and assessment.

M. Alejandra Sorto is an Associate Professor of Mathematics and Mathematics Education at Texas State

University. Dr Sorto holds a PhD in Mathematics Education from Michigan State University. Her

research interests include the development of instruments to assess teachers’ knowledge in mathematics

and statistics.

The effects of teacher mathematics knowledge and pedagogy 197

123