- Open Access
Exploring variation in measurement as a foundation for statistical thinking in the elementary school
International Journal of STEM Education volume 2, Article number: 3 (2015)
This study was based on the premise that variation is the foundation of statistics and statistical investigations. The study followed the development of fourth-grade students' understanding of variation through participation in a sequence of two lessons based on measurement. In the first lesson all students measured the arm span of one student, revealing pathways students follow in developing understanding of variation and linear measurement (related to research question 1). In the second lesson each student's arm span was measured once, introducing a different aspect of variation for students to observe and contrast. From this second lesson, students' development of the ability to compare their representations for the two scenarios and explain differences in terms of variation was explored (research question 2). Students' documentation, in both workbook and software formats, enabled us to monitor their engagement and identify their increasing appreciation of the need to observe, represent, and contrast the variation in the data. Following the lessons, a written student assessment was used for judging retention of understanding of variation developed through the lessons and the degree of transfer of understanding to a different scenario (research question 3).
The results were based either on the application of the hierarchical SOLO model or on non-hierarchical clustering of responses to individual questions in the student workbooks. Students' progress throughout the lessons displayed a wide range of explanations for the estimate of a single student's arm span, general surprise at the variation in measurements, and a large variety of hand-drawn representations based on the values or frequencies of measurements. Many different representations were also created in the software for the single student measurements and for the comparison of measurements for the two scenarios. Although the students' interpretations of their plots were generally more basic than sophisticated, the results of the assessment indicated that many students had developed the ability to transfer their appreciation of variation to another context and could clearly explain the meaning of variation.
The findings highlight the importance of an early focus on variation and distribution, with meaningful activities that motivate students to conduct and observe measurements, together with creating both hand-drawn and software representations to relate their experiences.
Since the introduction of statistics into school mathematics curricula about 25 years ago (e.g. Australian Education Council 1991; National Council of Teachers of Mathematics, 1989), there has been a growing awareness of the inadequacy of focusing solely on a procedural ability to calculate statistics. In particular, informal inference, a precursor to formal inference, has been highlighted as a foundational component that has not received the required attention especially in the elementary grades (Makar, in press). As described by Makar and Rubin (2009), informal inference is the process of using the evidence provided by data to answer questions beyond the data, acknowledging the uncertainty associated with the conclusion reached. Variation is the key to accepting a conclusion with some degree of certainty (Franklin et al., 2007, p. 18).
Variation is in fact such a fundamental concept that generally in the literature it is taken as an undefined term, as does Moore (1990). Variation and its associated form variability, however, are complex words for fourth-grade students and need to be described in ways that allow students to develop an understanding of the concept. Variation emerges when young students collect data and all of the values are not the same or when they plot the data and see that there are changes in the shape of the plot with gaps between and clusters of points/values. At a basic level, statistical variation may be regarded as the difference in data values and spread seen in data representations. Once variation is acknowledged, however, the purpose of a statistical investigation is to understand it, represent it, summarise it and perhaps control it (Franklin et al., 2007; Moore, 1990).
Acknowledging the importance of variation in informal inference, this study sought to expose fourth-grade students to the concept of variation through a measurement activity that would reveal two different types of variation: variation as ‘error’ in the repeated measurement of a single student's arm span and variation as the difference observed in single measurements of arm spans of a class of students. In the first scenario, all students measured one student's arm span where the sources of variation did not lie in the entity itself (one arm span), rather in the measurers, the measurement processes and the instruments used. In the second scenario, variation arose from the entities being measured (all students' arm spans), together with the measurers, their measurement processes and instruments. Students' documentation of their measurements, their explanations of the variations they observed, and their hand-drawn and software representations displaying the variations formed the data for addressing our research questions.
In the first scenario, we primarily explored students' levels of development in detecting, representing (by hand and with technology) and explaining variation in linear measurement values. In the second scenario, our major research focus was on the plots students created with software and their ability to compare and explain differences between their representations in the two scenarios. Finally, to determine the success of the activity in terms of students' understanding of variation and their ability to apply this learning to a slightly different scenario, we administered an assessment 1 to 3 months after activity completion. Specifically, our research questions were:
1. What are students' levels of development in detecting, representing and explaining variation in linear measurement values?
2. How do students develop the ability to compare and explain differences between representations in the two scenarios?
3. Do students demonstrate understanding of variation and transfer of that understanding through delayed assessment questions?
In addressing the background literature for the study, we consider the variation component of statistical thinking and inference, including its presence in national curriculum documents. We then examine the role of variation in linear measurement and conclude with a review of the emergence of variation in statistical representation.
Variation lies at the heart of statistical reasoning - without it, statistics would not exist (Moore, 1990). Variation is linked to all aspects of statistical investigations (Cobb and Moore, 1997; Garfield and Ben-Zvi, 2007; Konold and Pollatsek 2002; Watson, 2006). Indeed, as Watson (2006) indicated, the reason data are collected, graphs are created and averages are computed is to ‘manage variation and draw conclusions in relation to questions based on phenomena that vary’ (p. 21). Variation as the foundation of statistics provides the basis from which inferences can be drawn during statistical decision making. The confidence with which one can make a decision, however, depends on creating a balance between variation and expectation (Watson, 2005). This balance is expressed in tests of significance or confidence intervals in senior secondary school courses, but learning to appreciate variation and its relationship to expectation (in this study prediction) needs to begin very early with appropriate hands-on experiences and student/teacher questioning.
The explicit recognition of variation in curriculum documents has been slow, but recent publications have begun to mention the concept. Most notable among these is the GAISE report (Franklin et al., 2007) prepared under the auspices of the American Statistical Association. The report presents a four-stage framework for carrying out statistical investigations, with every stage including a focus on variation. Formulating Questions anticipates variation, Collecting Data designs for variation, Analysing Data accounts for variation in distributions and Interpreting Results allows for variation beyond the data. Such an approach does not underpin national curriculum documents to the same extent. In Australia, the curriculum refers to variation in the general description for learning up to second grade, in terms of developing an awareness of variation in data but specific mention does not appear until third grade (Australian Curriculum, Assessment and Reporting Authority 2013b). In New Zealand, the curriculum notes ‘acknowledging samples that vary’ at level 3, and variables are introduced at level 4, with consideration of sources of variation as integral to planning a statistical investigation at level 5 (Ministry of Education, 2009). In the United States, the Common Core State Standards Initiative (CCSSI) (2010)) introduces variation in the sixth grade, the first year where a concentrated focus on Probability and Statistics is included in its curriculum. These curriculum recommendations are the result of earlier research and provide the motivation for further research.
There has been substantial research on older students' and adults' awareness of variation (e.g. Makar and Confrey, 2005; Reading, 2004; Wild and Pfannkuch, 1999) but less so on how this understanding can be developed with young children. As numerous researchers have lamented, this neglect is a major concern especially given that secondary school and university students are usually ‘drilled’ in the use of statistical techniques without appreciating or understanding why, when or how these are applied sensibly to a range of situations (Bakker, 2004; Garfield and Ben-Zvi, 2005; Reading, 2004). The sentiments of Pfannkuch (2005) encapsulate these concerns: ‘Teaching has not focused on enculturating students into variation-type thinking and reasoning approaches. Much needs to be learnt about how to cultivate and develop this type of reasoning in students’ (p. 84).
The studies on young learners' awareness of variation have indicated the presence of a primitive understanding that needs to be integrated, revisited and emphasised in statistics learning from the earliest grade levels (Garfield and Ben-Zvi, 2007; Lehrer and Kim, 2009; Lehrer et al., 2011; Watson and Kelly, 2002a, 2002b). The research-based classroom suggestions from these studies, however, are not being adopted in many classrooms because teachers lack the pedagogical content knowledge to make specific links to variation when they implement activities in data and chance (Watson, 2013). Explicit discussion of variation and its application across the curriculum is needed throughout the elementary school years, before students meet formal measures such as standard deviation in the secondary school years. Instructional approaches to developing an understanding of variation in the earlier grades have received limited attention. As Garfield and Ben-Zvi (2005) noted, designing a long-term program that will nurture this understanding is a complex and challenging endeavour. Their suggestions for intervention were considered in planning the present classroom activity.
Assessing the overall outcomes of instructional approaches has also received limited attention. Watson et al., (2003) developed a survey instrument in the context of a study involving lessons emphasising variation in grade levels 3, 5, 7, 9 (Watson and Kelly, 2002a, 2002b). The outcomes suggested four levels of increasing appreciation of variation: ‘prerequisites,’ including interpreting the context, simple graph reading and intuitive chance reasoning; ‘partial recognition,’ including putting ideas in context but focusing on single aspects and neglecting others; ‘applications,’ including consolidating ideas in context but inconsistency on the most salient features; and ‘critical aspects,’ employing complex justification or critical reasoning. Although assessment in the present study was more general than that of the Watson et al. survey, their research suggests levels of development that are useful for assessing learning outcomes.
Variation in linear measurement
In basic terms, linear measurement may be viewed as ‘a comparative property of objects that embodies the amount of one-dimensional space between endpoints of the objects, which can be compared or quantified’ (Szilagyi et al., 2013, p. 583). Drawing on the literature, Szilagyi et al. cited numerous competencies needed for length measurement, including awareness of attribute, unit-attribute relations, unit iteration, proportionality and the relation between number and measurement. Transitivity and conservation are also important features. Apart from Lehrer's research (e.g. Lehrer et al., 2011), which we subsequently address, the focus on the important role of statistical foundations in the development of understanding of length measurement has been largely neglected.
The lack of measurement connectivity across other strands of the mathematics curriculum has been lamented by Australian educators (Lowrie et al., 2012) with respect to the recently implemented curriculum, The Australian Curriculum: Mathematics (Australian Curriculum, Assessment and Reporting Authority 2013b). Their concerns reflect repeated calls for more links within and across topics and disciplines with similar conceptual underpinnings (e.g. Bobis et al., 2009). Although measurement understandings have been linked to the development of geometry, number and algebra (e.g. Booker and Windsor, 2010; Bragg and Outhred, 2004), few studies have addressed connections with statistical literacy. Lehrer's research (e.g. Lehrer, 2007; Lehrer and Kim, 2009; Lehrer et al., 2011), however, has used measurement in developing students' appreciation of measures of centre and of variability. As an initial step in developing statistical reasoning, Lehrer and his colleagues engaged students in generating variability through measurement error. The findings of their research highlighted the importance of student interactions, where they made visible their arguments for a measure's validity and considered situations where such measures might be insufficient.
Despite Lehrer's work, little attention has been given to variation in measurement in curriculum documents and professional mathematics teaching articles focusing on length (e.g. Clements and Bright, 2003). Although the word ‘variation’ appears in the Australian, New Zealand and United States mathematics curricula (Australian Curriculum, Assessment and Reporting Authority 2013b; Common Core State Standards Initiative CCSSI 2010; Ministry of Education, 2009), it does not appear in relation to measurement. This is surprising given that variation is inherent in the entire measurement process - in the entity being measured, how the measurer undertakes the measuring process, the instruments used and how the measurements are interpreted. As a core understanding in the development of statistical literacy, variation includes interpreting and critically assessing various forms of measurement data. One of the goals of the present activity was the development of an appreciation of variation in measuring and measurements, and of the need for accuracy in measurement. Children need to understand what it means to make an accurate measurement, why accuracy is important, and the variation that can be expected in a measurement, especially if it is repeated (Lehrer, 2007; Watson and Wright, 2008).
The last understanding, of accuracy of measurement, is rarely addressed in the elementary curriculum yet as Konold and Pollatsek (2002) emphasised, it is an important context for various interpretations of average, an interpretation they refer to as ‘signal in noise’ (p. 268). From this perspective, each measurement is an estimate of an unknown yet specific value. We argue that connecting statistical and measurement topics can provide a powerful tool for targeting these currently neglected core understandings in the elementary curriculum and for addressing the long-standing concerns regarding students' difficulties with measurement (Smith et al., 2011).
Representing data in the elementary school years is typically limited to a few traditional graph types considered manageable by young students, such as picture graphs and simple column graphs, although recently using technology to produce these graphs has been suggested (see Australian Curriculum, Assessment and Reporting Authority 2013b, year 3). The focus has been largely on what types of graphs should be taught and when (e.g. Friel et al., 2001) and on appropriate scaling, labelling and directionality of plots (e.g. Leinhardt et al., 1990; Mevarech and Kramarsky, 1997). The ability of young students to create a wide range of representations of their own choosing is frequently underestimated (English, 2012, 2014; Russell, 1991). With a greater variety of graphs and other displays appearing in the media and elsewhere (e.g. Wall and Benson, 2009), students are being exposed more than ever to statistical representations that must be interpreted critically, where questions should be asked of the reliability of any claims made. This exposure is especially prevalent with the increased availability of technology, where students might create a range of representations without giving consideration to their appropriateness for the data being used (Konold, 2007).
Until recently, the explicit consideration of variation in relation to graphical representations has not been a feature of research. Yet, a key component of students' long-term development of statistical literacy is being able to interpret a message from a context with a distribution that displays variation, clustering, middles and unexpected values; this might not involve conventional text-book types of graphs (Lehrer, 2007; Wall and Benson, 2009; Watson et al., 2008). Early experiences with a range of representations that effectively display variation in data sets are important but have remained largely neglected in many elementary curricula until recently. Greater insights are needed into how young learners deal with variation in an immersive classroom environment including across disciplines (Watson, 2012), how they represent data in new ways of their own choosing, how they answer questions on comparing variation in different data sets and how they identify and justify the sources of variation they encounter.
In summary, this study addresses three underrepresented components of statistical development in the elementary grades. First, in integrating the strands of statistics and measurement in the curriculum, the study targets the neglected factor of variation in linear measurement. Second, the study highlights the importance of students' awareness, creation and understanding of variation in data representations, in contrast to traditional approaches where ‘information’ is simply read from a graph without being interpreted, analysed and questioned. Third, it focuses on an understanding and appreciation of informal inference, in this instance related to predicting a student's arm span length from data exhibiting variation, which is critical in interpreting data and distributions and making informed decisions.
Four year 4 (fourth grade) classes and one year 4/5 class from a large, middle socio-economic school participated during the first year of a 3-year longitudinal study (2012 to 2014). We focus only on the year 4 students (N = 115; mean age = 9.5 years) in the reporting of our findings, 43% of whom were classified as learners with English as their second language (ESL).
The 3-year longitudinal study was conducted from 2012 to 2014. The study took a form of design research involving the three phases of (i) preparation and design of instructional materials, (ii) teaching experiments and (iii) retrospective analyses (Cobb et al., in press; Cobb et al., 2003). These phases were conducted in each year of the study, with subsequent years being informed by the outcomes of the previous. The teachers had input and took primary responsibility for implementing the instructional program within their own classrooms. In this article, we only consider one comprehensive activity from the first year of the study, although the learning progression we developed across the 3 years was informed by the outcomes of each successive activity in which the students were engaged.
Preparation phase: designing the activities
In the first year of the study we designed, in conjunction with the teachers' input, three sets of activities, namely, a preliminary benchmarking activity where the students created, implemented and analysed a playground survey (English and Watson, in press). The purpose of this benchmarking activity was to provide the research team with starting points for the main activities for the beginning year. This was followed by the measurement activity (‘Measuring a Person's Arm Span’), which is the focus of this article. In the latter part of the first year, we implemented a multi-lesson probability activity that explored variation and expectation (English and Watson, in press).
In developing the activities, we drew upon a range of research but focused our attention primarily on the components of Garfield and Ben-Zvi's (2005) epistemological model for developing an in-depth understanding of variation, together with recommendations from the work of Lehrer and his colleagues (Lehrer, 2007; Lehrer et al., 2011), and previous positive outcomes with students in years 3 and 5 (Watson and Kelly, 2002a, 2002b). In particular, the suggestions of Garfield and Ben-Zvi were useful. Although their key ideas in each of the seven areas of knowledge of variability cover all years of schooling, the first three were especially relevant to our study, as was the last: (i) developing intuitive ideas of variation; (ii) describing and representing variability; (iii) using variability to make comparisons and (vii) considering variability as part of statistical thinking. These key ideas were the foundation of our planning for the activity conducted in year 4.
It was important in designing our program that it aligned with the school curriculum to ensure that the activities formed a core component rather than ‘add-on’ learning experiences. The teacher preparatory meetings conducted for each activity were thus vital in informing our design, as well as in enabling us to provide important professional development on the core statistical ideas to be developed. The present activity was designed in line with the Australian Curriculum: Mathematics (Australian Curriculum, Assessment and Reporting Authority 2013b), where the year 4 measurement strand states that students should ‘Use scaled instruments to measure and compare lengths’ (p. 43). For the data strand in year 4, the curriculum states that students should ‘Select and trial methods for data collection … Construct suitable data displays, with and without the use of digital technologies … Include tables, column graphs … Evaluate the effectiveness of different displays in illustrating data features including variability’ (p. 45). Because one of the general capabilities of the Australian curriculum is Information and Communication Technology (ICT), which includes ‘Investigation, Creating and Communicating with ICT’ (Australian Curriculum, Assessment and Reporting Authority 2013a, p. 53), the software TinkerPlots (Konold and Miller, 2011) was introduced as the digital technology to ‘construct suitable data displays’. The school was operating under this curriculum through the state education authority, where column graphs and picture graphs were specifically mentioned for year 4.
We targeted numerous learning objectives in designing the activity, not all of which are addressed in the research we report here. These objectives included developing students' appreciation of the need for accuracy in measurement and to pay careful attention to scale, competence in representing variation in measurement values and describing the shapes of data sets, confidence in predicting a representative measurement and skills in explaining variation in data sets. Learning to create various data displays in TinkerPlots and identify which displays best show the variation in a data set were also key objectives. An important learning feature was students' consideration of the most effective displays for showing the variation in the data sets, with the emerging understanding that there is very likely to be measurement error in the first scenario and, hence, the measurement in the second scenario is an approximation.
The creation of the activity involved several design iterations to ensure each component met our learning objectives and would generate adequate and appropriate data to gain insights into the students' learning. The authors and the senior research assistant worked on refining the activity, seeking teacher feedback as needed. Comprehensive teaching notes and corresponding student workbooks were developed, the latter serving as a core data source when completed by students.
Teaching experiment phase: implementing the activity
The ‘Measuring a Person's Arm Span’ activity comprised an initial teacher-led introduction to the activity, followed by student group work that was interspersed with whole class discussions on the groups' observations, their data representations and their findings. Implementation of the activity varied in time allocation per class, with an average duration of 5 h 10 min, spread across 3 days during 1 week for each class. Students worked in small groups (usually two or occasionally three), which were selected by the teachers based on students' ability to work collaboratively and communicate their thinking to others. The groups were of mixed achievement levels.
The teacher commenced the activity with a class discussion on how accurately ‘we’ can measure, different units of length we might use, and on ways length can be measured. The students were asked, ‘If we all measure the same object will we get the same answer for its length?’ On discussing issues of accuracy in this situation it was explained for the first scenario that, ‘We are all going to measure the arm span of the same person and see if we get the same value. We will consider how to display and compare these values and see if we can make a “best guess” as to the person's arm span. We will then discuss how confident we feel about having exactly the right value.’
Once all the measurements had been recorded, the students were asked to suggest other methods of organising the data other than the list on the whiteboard. The students then completed three questions in their workbooks about the recorded measurement values, reflecting Garfield and Ben-Zvi's (2005) development of intuitive ideas of variability and describing variability (phase (i)):
Were all of the values the same? Why or why not?
Were you surprised at some of the values? Which ones? Why?
Write a summary of how accurate you think the measurements in the table are. What is your ‘best guess’ of the arm span of the person the class measured? How confident are you of this value?’
The language of the classroom often translated ‘best guess’ to prediction, a difficult word for some of the ESL students.
The students subsequently discussed in their groups different ways to create a representation that they considered ‘best tells the story of the class measurements.’ No direction was given as to what type of representation to construct. Each student, however, created his/her own representation in the workbook. The next question asked students to ‘Write a summary statement about what your representation shows about the measurements your class made of the arm span of the person you measured’. The students were reminded to ‘Think about the variation that is seen in your plot or picture,’ reflecting Garfield and Ben-Zvi's (2005) ability to represent and describe variability (phase (ii)).
A whole-class stacked dot plot was then created on the white board. Students recorded their names and measured arm span values on PostIt notes and placed them in the correct positions on a horizontal axis drawn on the white board. The students were asked about the variation in values they observed and what they might expect if the arm spans of all students in the class were measured, as a prelude to the second measurement scenario. Prior to this second scenario, the students created their own dot plots of the single student's arm span measurements using the TinkerPlots software program. They were to include a textbox that explained the features of their graph, noting any clusters, gaps and outliers that might have occurred, again reinforcing the representing and describing of variation (Garfield and Ben-Zvi, 2005).
In the second scenario, the students were to measure and plot the arm span of all members of the class and record these measurements on the whiteboard. Due to time limitations, in most instances the researchers had to assist in taking these measurements. Prior to the measurements being taken, the students recorded their responses to two questions, namely:
Do you think all the values will be the same? Why/why not?
How accurate do you expect your results to be compared to our last lesson?
After recording their measurements on the white board, the student groups entered the data into a new data collection in their TinkerPlots files, which still contained their plots of the single student's arm span measurements. They were then to create a plot ‘that best describes the [new] data set and tells the story.’ This was to be accompanied by a text box containing a summary statement of what the plot shows and ‘at least two sentences that describe the differences between this plot and the earlier one,’ reflecting Garfield and Ben-Zvi's (2005) component of using variability to make comparisons (phase (iii)).
The activity concluded with a class discussion that highlighted the differences between the two scenarios and the variation in the two plots. The students were asked:
Why is there a difference in the spread/range for the two activities (scenarios)?
How confident are we of the values that we plotted in the second graph for each class member, having experienced the variation we found in the earlier plot for a single person?
Would it have been better to have made two or three or four measurements on each person and picked a ‘typical’ one?
These questions reflected Garfield and Ben-Zvi's (2005) last component of engendering consideration of variability in all parts of a statistical investigation (phase (vii)).
The data to be used in the retrospective analyses were drawn from multiple sources. In each class, we selected two focus groups for in-depth analysis, where their work was video- and audio-recorded and subsequently transcribed. All whole class discussions were also recorded and transcribed, and all of the students' workbooks were scanned for analysis. Between 1 and 3 months after completion of the activity, an assessment was implemented in all classes except the year 4/5 class whose timetable prevented their completion of the assessment. Two of the classes had 30 students undertake the assessment a month after completing the activity. Of these students, many were classified as ESL and/or low-achieving. Of the other two classes, streamed as high achieving, 36 students completed the assessment after 3 months. Although some were classified as ESL, many of these had high levels of English. The teachers of the two classes with the longer delay did not review the activity with the students before completing the assessment. Although there was a difference in performance between the two groups of two classes, favouring the class with the longer delay, it was believed to be the result of streaming the classes, and the data were combined for analysis of all year 4 students. The work of any students without parental consent was not recorded and not included in the analysis.
The data addressed here are drawn from the students' completed workbooks, their TinkerPlots creations and accompanying textbox descriptions, the transcriptions of whole class and focus group discussions and the assessment. Codes were checked by each author and refined, with the senior research assistant undertaking a final overview; consensus was reached on all coding.
In analysing the responses of students in their workbooks and in creating plots in TinkerPlots with accompanying text in text boxes, two kinds of criteria were used. In one situation, students were providing descriptions or reasons associated with observations or conjectures. These differed somewhat between classes because of the data collected or the way the teacher handled the discussion. It was of interest to categorise these responses to gain an appreciation of the diversity of thinking in the measurement context, but the responses were not considered to be hierarchical in nature. For these data, we employed content analysis (Patton, 2002). A description of the categories and the percentages of responses in each are presented in tables in the ‘Results’ section. In other situations where students were constructing representations or more complex arguments, it was possible to suggest a hierarchical cognitive structure reflecting increasing use of the elements available for the response. The criteria for levels of response are reflective of levels of the SOLO model (Biggs and Collis, 1982) where code 1 generally is assigned to employing a single element or aspect from the context of the question (a unistructural response). At code 2, several elements may be used, usually presented in sequence (a multistructural response) and, at code 3, relationships are demonstrated among the elements (a relational response). The particular descriptions of these levels where applicable are presented in the tables in the Results section.
The assessment began with straightforward questions based on a screen dump from the software for two reasons. The first was to confirm the students had picked up the basics of interpreting the information in the TinkerPlots format, and the second was to provide easy questions at the start of the assessment. The coding values for the instrument are given in the Additional file 1 with the questions asked, along with the rubric provided to teachers. Of interest is the total score obtained as it reflects the overall understanding of students in adapting to TinkerPlots as a context for reaching the learning objectives of the measurement activity. Further, the responses to questions 10 to 15, including a request for the definition of variation, throw light on the degree of application of ideas associated with variation in the classroom scenarios, as well as in a related context.
Although this study was not accompanied by extensive surveying that could lead to the type of Rasch analysis reported by Watson et al. (2003), it is possible to consider descriptively whether the responses provided by students, first in their workbooks, second in class discussions, and later in a limited assessment situation, exhibited the type of thinking typical of the four levels of their developmental hierarchy of understanding variation. Because of the specific scenarios of the classroom activity and the interaction of the teacher and students, it is assumed that the students understood the measurement environment for the activity. Although some were classified as ESL students with little English, there was no indication from the video of the classroom that any student was unaware of the task of measuring one student's arm span. They had also created graphs through previous classroom experiences, and it was assumed that all students had reached level 1 of the developmental hierarchy, Prerequisites for Variation.
We report the results in terms of the three research questions cited previously, namely, (1) What are students' levels of development in detecting, representing and explaining variation in linear measurement values? (2) How do students develop the ability to compare and explain differences between representations in the two scenarios? (3) Do students demonstrate understanding of variation and transfer of that understanding through delayed assessment questions?
The first research question is based on variation in linear measurement of one student's arm span. This links directly to level 2 of the developmental hierarchy, Partial Recognition of Variation, which involves putting ideas in context and focusing on a single element. The context is the consideration of variation in length measurements, repeated for a single student. The second research question is related to students' understanding of comparing variation in the two scenarios and ability to represent it in a visual way and then explain the meaning. This is one of the types of consolidation of ideas expected at level 3 of the developmental hierarchy, Application of Variation, where perhaps not all significant features are recognised. The third research question is related to critically interpreting the variation seen within and between the data distributions and transferring this to another context. This is the sort of complex justification expected at level 4 of the developmental hierarchy, Critical Aspects of Variation.
In reporting the results of the study in relation to the research questions, we first consider the outcomes from the classroom that are associated with the first two research questions. The third research question is addressed by the post-activity questions in the class assessment.
Research question 1: in the first scenario, what are students' levels of development in detecting, representing, and explaining variation in linear measurement values?
To determine the students' detection of variation and accuracy in their measurement values for the first scenario (measuring just one designated person in the class), we analysed their responses from the first three workbook questions (some of these responses were reported briefly in a conference paper of English and Watson, 2013). We identified different non-hierarchical categories of responses to each workbook question, with some students' responses encompassing more than one category, and some responses incomplete; hence the number of responses reported varies across the questions.
Workbook question 1: were all of the values the same? Why or why not?
Five main categories of responses were identified in the analysis of the first question. Although some responses were vague in presenting justifications for the different values, most students across the classes focused on the measuring tools and how they were used or their accuracy. Fewer students suggested measuring in different units, or the fact that the student being measured might have moved, as a reason for different values. The percentages for each category are given in Table 1.
Workbook question 2: were you surprised at some of the values? Which ones? Why?
Analysis of the children's responses (N = 95) to the question about surprising values yielded four categories (in addition to a category of four irrelevant or non-interpretable responses). Overwhelmingly, students identified an outlier or extreme/unusual value and included why it had occurred. Some students mentioned variation without focusing on a particular value, whereas others, especially in one class where there were no apparent outliers, noted the lack of variation, hence lack of surprise. The percentages are given in Table 2.
Workbook question 3: write a summary of how accurate you think the measurements in the table are. What is your ‘best guess’ of the arm span of the person the class measured? How confident are you of this value?
In analysing the students' assessment of the accuracy of their measurements, 101 responses were classified into eight groupings. The most frequently occurring category contained responses related to a modal idea, although the term mode was not used. Some other students focused on the idea of middle in choosing their prediction. Outside of these categories, students mentioned variation in the data, visual intuitions and reasons expressed for earlier questions. Outcomes are shown in Table 3.
The students created representations in two ways for the data collected in the first scenario, namely, constructing a hand-drawn representation on a blank page in their workbooks and then later using the TinkerPlots software to represent their data. For both formats, the students were to record a summary statement about what their representations showed about the measurements, keeping in mind variation seen in the plot. First we consider the students' hand-drawn representations and the associated explanations.
Students' hand-drawn representations and summary statements
For the hand-drawn representations, of the 84 students who completed this task, 6 students created two representations, making a total of 90. Of the 90 representations, 7% could not be interpreted, 36% focused on the actual measured values (V) and 58% focused on the frequency (F) with which the measured values occurred. Within these two types (measured values and frequency), it was possible to rank the representations based on their statistical appropriateness. For representations of value, it was considered that unordered lists were the most difficult to interpret, followed by unordered value plots with scales beginning at 0 or 10 (example V2 in Figure 1). Plots with scales beginning with a value of 80 or more (example V3 in Figure 1) were considered easier to interpret visually, whereas an ordered list was considered the most interpretable of representations of values.
For frequency representations, again order was considered more appropriate than non-ordered data as was a graphical presentation over tallies. Several of the plots appeared to represent frequency but were impossible to interpret (F0). Of the unordered representations, those based on tallies (example F1 in Figure 1) were considered more difficult to understand than those displaying frequency visually (example F2 in Figure 1). Of the representations that ordered the data, those based on tallies or totals (example F3 in Figure 1) were considered more difficult to interpret than those displaying frequency in plots. Further, the plots showing individual frequencies (example F5 in Figure 1), were thought more interpretable than those using intervals. Percentages of all levels are given in Table 4, with examples of some of the levels presented in Figure 1.
In writing a summary statement of what their hand-drawn representation conveyed about their data, of the 84 responses, some were not actually related to the data, such as mentioning colour or describing people. It was possible to employ a hierarchical coding of the responses linked to the SOLO model, with responses not related to the representation considered pre-structural and coded 0. A single acknowledgement of the graph as a focus was considered unistructural and coded 1. Putting together a description of either variation or expectation linked to the graph was considered multistructural (code 2), whereas linking both variation and expectation to the plot was considered relational (code 3) as this included the fundamental ingredients of a statistical analysis in making an ‘informal inference’ acknowledging variation. The most frequent response was to provide only a general description of the representation rather than its message. Table 5 reports the percentages of responses at each level with examples.
Students' TinkerPlots representations and summary statements
Students worked in pairs to create TinkerPlots graphs for the measurement of a single student. Thirty-six groups created plots and text boxes for this task. The codes for levels of understanding displayed again reflected the SOLO hierarchy with some groups not providing interpretable text (code 0), some groups focusing on single isolated aspects of their plots (code 1), other groups describing sequentially several aspects of their plots appropriately (code 2) and one group going further to relate the information in an argument (code 3). Descriptions and the percentages of groups with each code are given in Table 6. One or two examples of each code are shown in Figure 2.
Research question 2: how do students develop the ability to compare and explain differences between representations in the two scenarios?
Rather than create hand-drawn graphs of class data, students were asked to use TinkerPlots to compare and contrast the two data sets for the single student and the class. Thirty-one pairs of students produced plots and text boxes for this part of the activity. The hierarchical coding levels are listed in Table 7, with code 0 files not addressing the question or including any further advance on the previous question. At code 1, groups gave evidence of considering the class data only, not comparing it to the individual student data, whereas at code 2, groups made basic statements comparing the plots for the two situations (individual student and class). At code 3, two or more of the comparison statements were linked to give reasons for the differences in the two plots. Examples for codes 1 to 3, including the plots and text, are shown in Figure 3.
Insights into the students' thinking in comparing the representations between scenarios were gleaned from the class and focus-group transcripts. For example, during a class discussion on the differences between the two plots, Neil explained, ‘Shanna's was based on the same person and the other one was based on different [people], and it would be different measurement,’ and ‘Um, Shanna's arm span there's not like as much options in measurement, you can just do like one type cause it's her. Lots of people measure her and some people could get a not exact [measurement]’. In further discussion, another student explained why the two distributions differed, namely, ‘Shanna's would be more squished because the numbers would be closer together cause it's the same person and in, at the class, we're all different so it's all further apart’.
Students' expectation of accuracy in the second scenario
In the second scenario, while collecting measurements of all class members, students were to respond to the workbook question, ‘How accurate do you expect your results to be compared to our last lesson?’ As the data from each class were genuine, two classes had values that would be classified as outliers, two classes showed a large degree of variation, but no ‘certain’ outliers, and one class had very consistent measurements.
Six levels of responses (N = 85) were identified for this question. Those with no reference to criteria were coded 0. Some responses made reference to the measurement task without further explanation (code 1), whereas most of the students drew connections with descriptions of events during the activity that could affect accuracy (code 2). The most frequent percentage of responses again mentioned the measuring tools and how they were used. Other responses featured the accuracy of the person measuring (because in several classes the researchers collected the data due to time constraints), the greater experience of the students, and rather nebulous comments on ‘different’ people measuring. There was only one response (code 3) that related the critical issue of the number of measurements made on the individual student compared to the single measurements on the class. Although this issue was discussed with the class before the class measurements occurred, and again in discussion following students' completion of their representations (as we illustrate next), there were other issues that attracted the attention of the students. The percentages for each level of response are given in Table 8.
Students' assessment of accuracy in second representation
Despite only one written response indicating an awareness of the role of repeated measurements in assessing accuracy between scenarios, there was evidence of this understanding in class discussion following students' completion of the two plots. Given the variation in values the students identified in the first representation, they were asked to consider their confidence level in the values displayed in their second representation. Viviane, for example, explained, ‘I don't think they're exactly [accurate] because with S, in S's arm span we got lots of answers of the whole class and if, I think if we measured each of the person [second scenario] lots of times we would get different answers.’ Inta added to this point by commenting that it ‘would be better to do it three times because it would be like, once it could be 142, then the next one could be a different number and then the next one…’ Another student further commented that, ‘If you had three goes you could see, if you had 142 the first time, then 142 the next time, a different number the third time, you could estimate the 142 is the most popular, then you could do it again and see if it is the same.’ Reference to this approach being ‘a good idea’ was emphasised by another student, explaining that, ‘because … the one that is the most [frequently occurring value] will be good and all the ones in the middle will be best and if one's too small and one's too high, you don't [use them], you'd usually use the bigger ones [more frequently occurring value/s].
Research question 3: do students demonstrate understanding of variation and transfer of that understanding through the assessment questions?
There were 26 marks possible on the assessment questions (Additional file 1). The distribution of scores for the 66 students who completed the questions is shown in Figure 4, where the median score is seen to be 18. The distribution is somewhat skewed, with a mean of 16.7 because of the three very low scores of students who spoke virtually no English. Without those three scores, the median becomes 19 and the mean becomes 17.5, which seems a reasonable reflection of the performance of the students. The teachers were pleased that half of their students had achieved scores of 69% or more.
In further evaluating the students' understanding of variation and distribution within and across data representations, we consider their responses to the assessment items Q10 through Q15 (Additional file 1).
In reviewing the results for Q10 (associated with the figure at the beginning of the Additional file 1), on describing the plot shown (‘What it tells you about the people in the data set’), the maximum score was 4 for providing a meaningful and accurate account of the information in the plot. Although 20% of students did not score on Q10, 32% achieved a score of 4. The mean score across the students was 2.3 with 70% of the students achieving half marks or better.
Students scoring 1 provided very limited information on the plot, for example, ‘The shape of the plot is like building in the path’ and ‘It tells me in the data set that the people are quite old.’ Students scoring 2 gave some interpretation of the shape of the plot, with a sequence of observations: ‘There are more in the middle and are kind of spreader (sic.) on the out sides’ and ‘There are seven people are taller than 170 cm, eight people who are shorter than 164 cm and there are nearly 50% of people in the centre’.
Scores of 3 reflected both the content and some limited comments on shape: ‘It tells you 30 people heights. It shows a lot of option and goes from 146 cm to 188 cm.’ ‘It tells me about the people that their heights are all different. Some people are small’. Students who scored 4 provided meaningful links to a summary of the data. These included ‘It shows the average is about 164 to 170 cm which is in the middle and it is irregular with random bumps’ and ‘The shape is a bit of a triangle. This means it starts small on the edges and goes up to centre. Centre = 164 to 170 14 people are in the centre. There are eight people around the edges. The average person is about 167 cm’.
Q11 aligned with the classroom activity, asking students to distinguish between arm span measurements for a single student (lower plot) and a class (upper plot) (see Additional file 1), followed by the reason for being able to distinguish correctly. Seventy-seven percent of responses correctly distinguished the plots, 3% had labelling difficulties, and 20% either did not answer or did not use the labels provided. For Q12 where students explained how they decided which plot was which, 20% of students could not answer the question but 62% scored the maximum of 4, reflecting a good understanding of the classroom activity, as indicated in the sample responses below for a score of 4.
The top one was the 19 students armspan because there was a lot of difference and the bottom one was Leo's armspan because it was all cramped up together. Leo's armspan didn't have a lot of variation because it is only one arm to measure and it should be nearly the same. While the 19 students one had alot of spread because its unlikely that one person could have a range of armspands [arm spans].
If Leo's plot was the top one it would be impossible because Leo can't have that many armspans. If it was 19 people everyone would most likely have a different range. The top one has a bigger variation.
Responses that were scored 2 marks (11%) only considered one plot for the reasoning, for example, ‘I know the second plot is Leo's Armspan because its bunch[ed] up in one place’ and ‘I think the bottom one is Leos (sic) because the scores are mostly the same and Leo only has one armspan.’
Q13 and Q14 were based on a plot that compared arm span measurements for two grades rather than one student and a class. Students were asked if there were an overall difference in the arm spans of these grade 6 and grade 7 students. The plot is shown above Q13 in the Additional file 1. Q13 asked for their answers and Q14 asked for two reasons for the decision. For Q13 (not scored), one student did not answer, 82% said ‘Yes’, 12% said ‘Not sure’ and 5% said ‘No’.
Two marks were possible for each reason in Q14, and there was an indication that students did slightly better for the first reason than the second. This is likely to have occurred because it is more difficult to think of two reasons than one. Overall, for both reasons, the mean score was 1.2. If a reason was repeated with little change, no mark was given for the second. The ‘not sure’ responses were generally followed by ‘I don't know’ or a blank. Some reasons for ‘Yes’ answers did not provide any specific information from the plot (score of 0), e.g. ‘Yes, because you can see it’ and ‘People have differnet [different] heights in diff[e]rent age groups.’ Scores of 1 generally only mentioned a single feature of the plot or a general recognition of the context: ‘If you look at the top of the graph there is one dot in the very end’ and ‘Yes because normally Year 7's have bigger armspans (sic.) and I've seen it.’ Responses that scored 2 marks displayed a general recognition of difference in shape, as in, ‘Because one is more to the right and one more to the left’ and ‘Because the graph shows that grade 6's are bunched up at the small end’. Two of the ‘No’ responses actually gave a reason for the classes to be different whereas the only valid reason supporting no difference was the following: ‘I gave my answer no because the difference between the armspan (sic.) is not big (it is medium not big).’
For the final question, Q15, asking for the meaning of variation (worth 2 marks), the mean score was 1.4 with 64% of students achieving a score of 2. Many of the 26% of responses that were scored 0 were of the form, ‘I don't know.’ Others reflected an understanding of the context but not the nature of variation, such as, ‘Long of armspan’ or ‘The number of people.’ Responses with a score of 1 often used a word with the same root as variation, for example, ‘I think it means varity [variety]’ and ‘Varying the amount by which something varies.’ The concept of difference among some things being measured was reflected in score 2 responses: ‘Variation means difference or many things, a variety for example, Sam's 131 cm and Jonah is 156 cm that's the variation’; ‘It means how different things are (comparing);’ and ‘The diff[e]rences or similaritys [similarities] with something’.
The mean score (out of 16) for the scored items Q10 to Q15 was 10.3, with a median of 12, suggesting that the students had developed a reasonable understanding of the core concepts that were targeted in the activity. In particular, the students appeared to have developed a basic understanding of variation (Q15), scoring a mean of 1.4 out of 2. This understanding was borne out in students' responses to Q10 where well over half the class could describe and interpret the shape of the plot of heights. Likewise, for Q11 and Q12, the majority of students could identify the correct labels for the two arm span plots and could explain why this was the case.
In reflecting on our findings, we address each research question in turn, beginning with the students' developments in detecting, representing and explaining variation in linear measurement values. We next consider their learning as they compared and explained differences between representations in the two measurement scenarios, including identifying variation and distributions, judging measurement accuracy and appreciating the role of repeated measurements. Finally, we review the students' retention and transfer of variation and distribution, including describing plots and their shape and identifying value clusters and outliers. In reflecting on our findings we touch upon the issue of how students conceive of variation and distribution, in terms of whether the former is seen as generating the latter or vice versa.
Detecting, representing and explaining variation in linear measurement values
With respect to the first research question (measuring one designated class member), the students recognised that the use of different measuring tools and how accurately they were used led to variation in values obtained and difficulty in predicting the student's actual arm span. Any outliers that occurred were readily detected and, if no such values existed, the students considered their cluster of values to predict the approximate arm span of the student. This reference to a modal notion was evident in the students' reflections on their accuracy, with recognition of the ‘most popular’ value/s as the likely arm span measurement. Variation in values, however, was considered by some to indicate less accuracy, as was the way in which the measuring tools were used. Although the language of informal inference was not used at this grade level, the students' best guesses or predictions were made with implicit acknowledgement of the uncertainty associated with the presence of variation.
The use of a real-world measurement activity, in which variation is a naturally occurring feature, provided a rich context for students' appreciation of phenomena that vary. As Lehrer's research has indicated (e.g. Lehrer et al., 2011), engaging young students in creating variability through measurement error is a powerful way of making visible the impreciseness in measuring - from the tools chosen, how they are manipulated and how measurements are recorded and interpreted. Experiences in undertaking repeated measurements are especially important here as they reveal the inherent variation more explicitly and convincingly. Furthermore, as observed by Lehrer et al. (2011), as students attempt to take repeated measures they became more aware of the phenomenon being measured and how any changes in that phenomenon would impact on the measurements generated. What the present activity provided was a hands-on experience of variation in a context designed to build beginning recognition of variation. In terms of the Watson et al. (2003) developmental hierarchy, this early identification of variation represents at least the second level of development for nearly all students, whereas many students reached the third level for this component of informal inference.
An awareness of the ‘affordances and constraints of measures in light of characteristics of distribution’ (Lehrer et al., 2011, p. 275) was becoming apparent as the students proceeded to represent and interpret their data by generating a range of representations, both hand-drawn and using the TinkerPlots software. Recognition of clusters of values was evident in their hand-drawn representations, with over half of the students focusing on the frequency with which the measured values occurred. Students' representations enabled them to gain a sense of the ‘signal in noise’ (Konold and Pollatsek, 2002, p. 268) in their data, that is, an indication of what the actual arm span measurement of the designated person might be.
Although the ordering of values and the creation of frequency plots with meaningful scales were prevalent in these hand-drawn representations, many (approximately one third) featured less sophisticated displays of actual measured values. Although our analysis focused on the statistical sophistication of the students' data displays rather than the type of representation, it is worth noting that a wide range of representations was produced. There was the use of lists, tally marks, bar graphs, circle graphs, dot plots, and histograms, with some students creating more than one representational form.
In contrast to this representational diversity were the somewhat limited descriptions of what the representations conveyed about the data. Nearly half of the responses provided only a general description of the representation rather than its message, typical of partial recognition of variation (level 2). Nevertheless, approximately a third of responses made reference to the variation and expectation that their hand-drawn representation revealed.
The students displayed similar responses in their text box explanations of their software representations, with a simple reporting of the plot features prevalent. Although there was some reference to data distribution features such as the range and most frequent values, it appeared that interpreting representations beyond a literal reading (cf. Curcio, 2010) was a new experience for the students and that more attention is required here, especially given young students' competence in generating a range of representational types (English, 2013).
Comparing and explaining differences between representations in two scenarios
The comparison of data from the two scenarios provided the opportunity for students to develop and display a consolidation of ideas related to variation in a realistic context. Identifying variation across these data sets represented development at level 3 of the Watson et al. (2003) hierarchy, although there were some inconsistencies in the presentation of arguments as evident in the students' written responses on accuracy. The class discussion transcripts, however, indicated some students could express in colloquial language ideas that they found difficult to express in written language in their workbooks.
The observed outcomes in the classroom suggested that a focus on developing students' ‘metarepresentational competence’ is needed, a term that diSessa and his colleagues (e.g. diSessa, 2004; diSessa et al., 1991) proposed to indicate the range of students' capabilities in constructing and using external representations prior to direct instruction - and even independently of it. This metarepresentational competence, however, needs to be accompanied by the development of conceptual competence (diSessa, 2004); specifically, students should be able to interpret and explain the statistical features displayed by their representations. As diSessa noted, research is limited here, and the role of student-created representations in conceptual development is complex. Further investigation is needed on how certain strengths or limits of metarepresentational competence might advance or hinder conceptual competence, especially as students move from non-software to software-generated representations. At the same time, it is necessary to address whether metarepresentational competence and conceptual competence grow together as students develop the statistical competencies we have explored here (English, 2013).
Although we cannot answer these questions directly from our results, students' comparisons of their software generated representations for the two measurement scenarios appeared to indicate concurrent conceptual and metarepresentational development. Over half of the student pairs could detect distribution differences, with some responses more sophisticated than others. Drawing analogies between familiar shapes (e.g. triangle) and the data distribution display was common, but reference to specific statistical information (e.g. clusters of values, variation) was not as prevalent.
In terms of judging measurement accuracy between the two scenarios, the majority of students recognised variation in the tools used and the reliability with which they were used. A few commented on the learning they had gained from the first scenario as a reason for improved accuracy in the second, suggesting a metacognitive awareness of the knowledge they had developed. This awareness has not received the attention it warrants, as both Lester (2013) and Schoenfeld (2013) have emphasised. Indeed, very little is known about how to develop students' metacognitive abilities even though they play a critical role in mathematical problem solving and learning in general.
The role of repeated measurements, the most appropriate statistical response to the question regarding measurement accuracy in the second scenario, was only mentioned once in the students' workbook responses. Follow-up class discussion, however, suggested some appreciation and understanding of the impact of repeated measurements. Very little attention has been given to this important component, which is critical to the development of long-term statistical reasoning (Cobb and Moore, 1997; Lehrer et al., 2011; Noll and Shaughnessy, 2012; Thompson et al., 2007). For elementary students to develop competency with measurement, they need to be exposed to experiences in which they witness, first-hand, the impact of repeated measurements and hence the variation that is inherent in processes of measurement.
Retention and transfer of variation and distribution
Students' understanding of variation and distribution within and across data representations was further gauged from the assessment items (third research question). In general, the students were able to describe the plots of the type they had created, noting the plot shape and identifying value clusters and outliers. On the one hand, the extent to which they provided meaningful links to a summary of the data, such as reference to centre or average or most, was not prevalent; however, we did not specifically introduce the formal notions of mean, median and mode, because these are not part of their fourth-grade curriculum. On the other hand, the majority of the students were able to distinguish between plots displaying measurements for a single student and for all students, indicating a sound understanding of the core ideas of the classroom activity.
For the assessment item that asked students to distinguish between two groups (a more general comparison than they had considered in class), 77% of students achieved a score of 1 or 2 for their first reason. This dropped to 68% for the second reason. Because the students had not seen a comparison like this before, it was felt that the performance indicated an encouraging degree of transfer of understanding of variation and the ability to think critically in a similar but non-identical context. Furthermore, well over half of the students could clearly explain the meaning of variation, demonstrating that this core concept had been absorbed meaningfully for these students.
Students' application of critical reasoning, typical of level 4 in the developmental hierarchy of Watson et al. (2003), was not as prevalent in their workbooks as we would have liked. This could have been due partially to the time constraints necessitating the researchers to assist in the measurements of the class arm spans. The class discussion, however, helped students make the desired connection with accuracy of measurement. Further, on the post-activity assessment questions, many students were able to define variation and demonstrate justifications typical of the level 4 developmental hierarchy.
These findings raise the issue of how students viewed variation and distribution. It could be that variation was seen as generating distribution, which Prodromou and Pratt (2006) considered to be half of the desired understanding. That is, this ‘data centric’ perspective needs to be complemented by a ‘modelling perspective’ where distribution is perceived as variation from an ideal model (pp. 70-71). Yet, as Noll and Shaughnessy (2012) noted, coordinating both perspectives can be challenging for young students especially when school curricula focus on one at the expense of the other. Watson's (2009) research on developing awareness of distribution, for example, suggested that students' natural intuition about variation could be overshadowed when curricula emphasise expectation. We thus argue that attention needs to be focused on both perspectives in order to sow the seeds for strong statistical reasoning. Clearly more research is warranted here including how students think about data distributions in different contexts (Noll and Shaughnessy, 2012). Because data are the basic building blocks for the study of distributions and their variation tells the stories therein, it is essential that students develop a meaningful appreciation that is deeper than numbers on a page or dots on a line. Although the language of distribution may be very complex for students to absorb in fourth grade alongside variation, slowly students should develop the notion suggested by Wild (2006) that distribution is ‘the lens through which statisticians look at the variation in data’ (p. 10).
Variation is the foundation of all of statistics and an underlying feature of all measurement. It is thus paramount that an early start is made in providing meaningful activities that are motivating both in their intrinsic interest for the students in carrying out and observing the measurements and in the challenge of creating representations to tell the story of their experiences. In terms of the mathematics curriculum (at least in Australia), the study demonstrates the advantage of linking internal sections of the curriculum in meaningful ways. As noted earlier, the measurement part of the mathematics curriculum has generally received diminished attention in recent years, and activities such as this can regenerate interest in measurement as well as contributing to the development of essential understandings in statistics. In terms of variation itself, the study reinforces the importance of the GAISE approach (Franklin et al., 2007) to variation in all stages of a statistical investigation. It is to be hoped that future versions of mathematics curricula reflect this way of thinking in setting the foundations for the statistics component of the curricula.
We have highlighted a number of issues in need of further research and ultimately inclusion in school curricula. These issues include more studies that utilise real-world situations where variation is a naturally occurring feature. In the present case, the creation of variability through measurement error provides a powerful means of making visible the impreciseness in measuring, an understanding that needs greater attention in the elementary school. More opportunities for students to make data comparisons, to create and interpret data representations and recognise their core distributional features and to read critically ‘beyond the data’ (cf. Curcio, 2010) are also in need of research and curriculum development. Part of this research is the underrepresented component of metarepresentational competence (e.g. diSessa 2004), which we have argued is particularly important in students' software generated representations. Finally, the issue raised earlier regarding coordinating the two perspectives on variation and distribution, that is, variation as generating distribution and distribution as variation from a desired ‘model’ (Prodromou and Pratt, 2006), warrants increased curriculum and research consideration.
In concluding, the limitations of the study warrant noting. Given that the study was conducted in a regular classroom setting, a number of usual obstacles were faced. These included the limited time available, the overcrowding of computer labs, last-minute timetable changes and unexpected events or changes of direction in classroom discussions. In follow-up feedback, however, all teachers were positive about their participation in the study and expressed surprise at the achievements of some students whom they did not consider would meet the expectations of the activity.
Australian Curriculum, Assessment and Reporting Authority. (2013a). General capabilities in the Australian Curriculum, January, 2013. Sydney, NSW: ACARA.
Australian Curriculum, Assessment and Reporting Authority. (2013b). The Australian curriculum: mathematics, version 5.0, 20 May 2013. Sydney: NSW: ACARA.
Australian Education Council. (1991). A national statement on mathematics for Australian schools. Melbourne: Author.
Bakker, A. (2004). Reasoning about shape as a pattern in variability. Statistics Education Research Journal, 3(2), 64–83.
Biggs, JB, & Collis, KF. (1982). Evaluating the quality of learning: the SOLO taxonomy. New York: Academic Press.
Bobis, J, Mulligan, J, & Lowrie, T. (2009). Mathematics for children: challenging children to think mathematically (3rd ed.). Frenchs Forest, NSW: Pearson Education.
Booker, G, & Windsor, W. (2010). Developing algebraic thinking: using problem solving to build from number and geometry in the elementary school to the ideas that underpin algebra in high school and beyond. Procedia – Social and Behavioral Sciences, 8, 411–419.
Bragg, P, & Outhred, L. (2004). A measure of rulers – the importance of units in a measure. In MJ Horne & AB Fuglestad (Eds.), Proceedings of the 28th International Conference of the International Group for the Psychology of Mathematics Education (PME) (Vol. 2, pp. 159–166). Bergen: PME.
Clements, DC, & Bright, G. (2003). Learning and teaching measurement: 2003 yearbook. Reston, VA: National Council of Teachers of Mathematics.
Cobb, GW, & Moore, DS. (1997). Mathematics, statistics, and teaching. American Mathematical Monthly, 104, 801–823.
Cobb, P, Confrey, J, diSessa, A, Lehrer, R, & Schauble, L. (2003). Design experiments in educational research. Educational Researcher, 32(1), 9–13.
Cobb, P, Jackson, K, & Munoz, C (in press). Design research: A critical analysis. In LD English & D Kirshner (Eds.), Handbook of international research in mathematics education (3rd ed.). New York: Routledge.
Common Core State Standards Initiative (CCSSI). (2010). Common core state standards for mathematics. Washington, DC: National Governors Association for Best Practices and the Council of Chief State School Officers. Available at: http://www.corestandards.org/assets/CCSSI_Math%20Standards.pdf.
Curcio, FR. (2010). Developing data-graph comprehension in Grades K-8 (3rd ed.). Reston, VA: National Council of Teachers of Mathematics.
diSessa, AA. (2004). Metarepresentation: native competence and targets for instruction. Cognition and Instruction, 22(3), 291–292.
diSessa, AA, Hammer, D, Sherin, B, & Kolpakowski, T. (1991). Inventing graphing: metarepresentational expertise in children. Journal of Mathematical Behavior, 10, 117–160.
English, LD. (2012). Young children's metarepresentational competence in data modelling. In J Dindyal, L Pien Cheng, & S Fong Ng (Eds.), Mathematics education: expanding horizons (Proceedings of the 35th annual conference of the Mathematics Education Research Group of Australasia, eBook (pp. 266–273). Singapore: MERGA.
English, LD. (2013). Reconceptualising statistical learning in the early years. In LD English & J Mulligan (Eds.), Reconceptualising early mathematics learning (pp. 67–82). Dordrecht: Springer.
English, LD. (2014). Promoting statistical literacy through data modelling in the early school years. In E Chernoff & B Sriraman (Eds.), Probabilistic thinking: presenting plural perspectives (pp. 441–457). Dordrecht: Springer.
English, LD, & Watson, J. (2013). Beginning inference in fourth grade: Exploring variation in measurement. In V Steinle, L Ball, & C Bardini (Eds.), Mathematics education: yesterday, today and tomorrow (Proceedings of the 36th annual conference of the Mathematics Education Research Group of Australasia) (pp. 274–281). Melbourne: MERGA.
English, LD, & Watson, JM. (in press). The development of probabilistic understanding in fourth grade. Journal for Research in Mathematics Education.
English, LD, & Watson, JM. (in press). Statistical literacy in the elementary school: Opportunities for problem posing. In F Singer, N Ellerton, & J Cai (Eds.), Problem posing: From research to effective practice. Dordrecht: Springer.
Franklin, C, Kader, G, Mewborn, D, Moreno, J, Peck, R, Perry, M, & Scheaffer, R. (2007). Guidelines for assessment and instruction in statistics education (GAISE) report: A preK-12 curriculum framework. Alexandria, VA: American Statistical Association. Retrieved July 3, 2009 from http://www.amstat.org/education/gaise/.
Friel, SN, Curcio, FR, & Bright, GW. (2001). Making sense of graphs: critical factors influencing comprehension and instructional implications. Journal for Research in Mathematics Education, 32, 124–158.
Garfield, J, & Ben-Zvi, D. (2005). A framework for teaching and assessing reasoning about variability. Statistics Education Research Journal, 4(1), 92–99.
Garfield, J, & Ben-Zvi, D. (2007). How students learn statistics revisited: a current review of research on teaching and learning statistics. International Statistical Review, 75(3), 372–396.
Konold, C. (2007). Designing a data analysis tool for learners. In MC Lovett & P Shah (Eds.), Thinking with data (pp. 267–291). New York: Lawrence Erlbaum.
Konold, C, & Miller, CD. (2011). TinkerPlots: dynamic data exploration [computer software, Version 2.0]. Emeryville, CA: Key Curriculum Press.
Konold, C, & Pollatsek, A. (2002). Data analysis as the search for signals in noisy processes. Journal for Research in Mathematics Education, 33(4), 259–289.
Lehrer, R. (2007). Introducing students to data representation and statistics. In J Watson & K Beswick (Eds.), Mathematics: essential research essential practice (Proceedings of the 30th Annual Conference of the Mathematics Education Research Group of Australasia) (Vol. 1, pp. 22–41). Adelaide: MERGA.
Lehrer, R, & Kim, M. (2009). Structuring variability by negotiating its measure. Mathematics Education Research Journal, 21(2), 116–133.
Lehrer, R, Kim, M, & Jones, RS. (2011). Developing conceptions of statistics by designing measures of distribution. ZDM Mathematics Education, 43, 723–736.
Leinhardt, G, Zaslavsky, O, & Stein, MK. (1990). Functions, graphs and graphing: tasks, learning and teaching. Review of Educational Research, 60(1), 1–64.
Lester, FK. (2013). Thoughts about research on mathematical problem-solving instruction. The Mathematics Enthusiast, 10(1&2), 245–278.
Lowrie, T, Logan, T, & Scriven, B. (2012). Perspectives in geometry and measurement in the Australian Curriculum: Mathematics. In B Atweh, M Goos, R Jorgensen, & D Siemon (Eds.), Engaging the Australian National Curriculum: Mathematics—Perspectives from the field (pp. 71–88). Online Publication: Mathematics Education Research Group of Australasia.
Makar, K (in press). Developing young children's emergent inferential practices in statistics. Mathematical Thinking and Learning.
Makar, K, & Confrey, J. (2005). Using distributions as statistical evidence in well-structured and ill-structured problems. In K Makar (Ed.), Proceedings of the Fourth International Research Forum on Statistical Reasoning, Thinking and Literacy (SRTL-4). SRTL4 - 2005: Fourth International Research Forum on Statistical Reasoning, Thinking and Literacy. New Zealand: University of Auckland.
Makar, K, & Rubin, A. (2009). A framework for thinking about informal statistical inference. Statistics Education Research Journal, 8(1), 82–105. www.iase-web.org/documents/SERJ/SERJ8(1)_Makar_Rubin.pdf.
Mevarech, ZR, & Kramarsky, B. (1997). From verbal descriptions to graphic representations: stability and change in students' alternative conceptions. Educational Studies in Mathematics, 32, 229–263.
Ministry of Education. (2009). The New Zealand curriculum: mathematics standards for years 1–8. Wellington, NZ: Author. Available at http://nzcurriculum.tki.org.nz/The-New-Zealand-Curriculum/Learning-areas/Mathematics-and-statistics.
Moore, DS. (1990). Uncertainty. In LS Steen (Ed.), On the shoulders of giants: new approaches to numeracy (pp. 95–137). Washington, DC: National Academy Press.
National Council of Teachers of Mathematics. (1989). Curriculum and evaluation standards for school mathematics. Reston, VA: Author.
Noll, J, & Shaughnessy, JM. (2012). Aspects of students' reasoning about variation in empirical sampling distributions. Journal for Research in Mathematics Education, 43, 509–556.
Patton, MQ. (2002). Qualitative research and evaluation methods (3rd ed.). Thousand Oaks, CA: Sage Publications Inc.
Pfannkuch, M. (2005). Thinking tools and variation. Statistics Education Research Journal, 4(1), 83–91.
Prodromou, T, & Pratt, D. (2006). The role of causality in the co-ordination of the two perspectives on distribution within a virtual simulation. Statistics Education Research Journal, 5(2), 69–88. http://iase-web.org/documents/SERJ/SERJ5%282%29_Prod_Pratt.pdf.
Reading, C. (2004). Student description of variation while working with weather data. Statistics Education Research Journal, 3(2), 84–105.
Russell, SJ. (1991). Counting noses and scary things: children construct their ideas about data. In D Vere-Jones (Ed.), Proceedings of the Third International Conference on the Teaching of Statistics (pp. 158–164). Dunedin, NZ: University of Otago.
Schoenfeld, AH. (2013). Reflections on problem solving theory and practice. The Mathematics Enthusiast, 10(1&2), 9–34.
Smith, JP, van den Heuvel-Panhuizen, M, & Teppo, A. (2011). Learning, teaching, and using measurement: introduction to the issue. ZDM Mathematics Education, 43, 617–620.
Szilagyi, J, Clements, DH, & Samara, J. (2013). Young children's understanding of measurement: evaluating a learning trajectory. Journal for Research in Mathematics Education, 44, 3.
Thompson, PW, Liu, Y, & Saldanha, LA. (2007). Intricacies of statistical inference and teachers' understanding of them. In MC Lovett & P Shah (Eds.), Thinking with data (pp. 207–231). New York: Lawrence Erlbaum.
Wall, JJ, & Benson, CC. (2009). So many graphs, so little time. Mathematics Teaching in the Middle School, 15, 82–91.
Watson, JM. (2005). Variation and expectation as foundations for the chance and data curriculum. In P Clarkson, A Downton, D Gronn, M Horne, A McDonough, R Pierce, & A Roche (Eds.), Building connections: theory, research and practice (Proceedings of the 28th annual conference of the Mathematics Education Research Group of Australasia, Melbourne (pp. 35–42). Sydney: MERGA.
Watson, JM. (2006). Statistical literacy at school: growth and goals. Mahwah, NJ: Lawrence Erlbaum.
Watson, JM. (2009). The influence of variation and expectation on the developing awareness of distribution. Statistics Education Research Journal, 8(1), 32–61.
Watson, JM. (2012). History and statistics: connections across the curriculum. AGORA, 47(3), 58–64.
Watson, JM. (2013). Statistical literacy, a statistics curriculum for school students, the pedagogical content needs of teachers, and the Australian Curriculum. Curriculum Perspectives, 33(3), 58–69.
Watson, JM, & Kelly, BA. (2002a). Can grade 3 students learn about variation? In B Phillips (Ed.), Proceedings of the Sixth International Conference on Teaching Statistics: developing a statistically literate society, Cape Town, South Africa. Voorburg, The Netherlands: International Statistical Institute. Retrieved from http://iase-web.org/documents/papers/icots6/2a1_wats.pdf.
Watson, JM, & Kelly, BA. (2002b). Grade 5 students' appreciation of variation. In A Cockburn & E Nardi (Eds.), Proceedings of the 26 th Conference of the International Group for the Psychology of Mathematics Education (Vol. 4, pp. 385–392). Norwich, UK: University of East Anglia.
Watson, J, & Wright, S. (2008). Building informal inference with TinkerPlots in a measurement context. Australian Mathematics Teacher, 64(4), 31–40.
Watson, JM, Kelly, BA, Callingham, RA, & Shaughnessy, JM. (2003). The measurement of school students' understanding of statistical variation. International Journal of Mathematical Education in Science and Technology, 34, 1–29.
Watson, JM, Fitzallen, NE, Wilson, KG, & Creed, JF. (2008). The representational value of hats. Mathematics Teaching in the Middle School, 14, 4–10.
Wild, CJ. (2006). The concept of distribution. Statistics Education Research Journal, 5(2), 10–26.
Wild, CJ, & Pfannkuch, M. (1999). Statistical thinking in empirical enquiry (with discussion). International Statistical Review, 67(3), 223–265.
The authors declare that they have no competing interests.
The authors shared equally in all parts of the study. Both authors read and approved the final manuscript.
About this article
Cite this article
English, L.D., Watson, J.M. Exploring variation in measurement as a foundation for statistical thinking in the elementary school. IJ STEM Ed 2, 3 (2015) doi:10.1186/s40594-015-0016-x
- Elementary school
- Graphical representations
- Statistical thinking