fbpx

So, first is a new recent press release in The Age - taken 30/07/2025.

After that are two articles on NAPLAN that I have produced some time ago. At the finish, I have collected a number of recent articles to compare with the relevance of my oldish work.

Victorian kids are the brightest sparks and NAPLAND downturn.

 The Age 30/07/2025

“Victoria’s schoolchildren have defied a national trend of lack lustre results in this year’s NAPLAN tests to catapult the state to the top of the class.

Nationwide NAPLAN results show little or no improvement on the previous year, with large numbers of children still struggling with the basics of learning and nearly a third of students in years 3, 5, 7 and 9 playing catch-up in reading, writing, spelling and maths.

The results of tests sat by more than 1.3 million children around the nation in March show nearly 33 per cent are either “in need of additional support” or “developing”, on average, across reading and writing.

The data also highlights the sharp economic and demographic divides in Australian education, with more than 57 per cent of Australia’s most disadvantaged year 9s – and about 53 per cent of Victoria’s – struggling with their reading, compared with just 17 per cent of those from the top socioeconomic bracket.

Despite the underwhelming national picture, Victorian students recorded either the best or second-best results in the country in 18 of the 20 categories reported, in what the state government describes as the “best ever” NAPLAN performance by local students.

Victoria also recorded fewer students struggling across all subjects and age groups, with less than 28 per cent of the state’s children needing additional support or “developing”, compared with a national average of almost 33 per cent.

The state government said the improvement was a direct result of its push towards an explicit-instruction approach to teaching maths and literacy, but a leading expert said Victoria could and should be doing better.

Education Minister Ben Carroll said there was a direct line between the Victorian government’s high-profile push for explicit learning and the improved results.

“The credit goes mostly to our hardworking teachers because, at the end of the day, pedagogy is the single most important lever we have in uplifting standards,” Carroll said. “So, explicit instruction is now mandated in Victorian teaching and learning models that I have launched, and that is what is occurring.

“I see it daily, whether it’s mathematics, whether it’s reading, our teachers are using what the science says is best practice, and it’s being now consistent across our classrooms and our schools.”

Jordana Hunter, education program director at progressive think tank the Grattan Institute, acknowledged the “green shoots” of Victoria’s results, but said the state’s schools had the potential to do much better.

“Victoria is a highly advantaged jurisdiction, so would you expect it to be doing better than Tasmania or South Australia or the Northern Territory? Absolutely,” Hunter said.

“When we look internationally, for kids in primary school, for maths outcomes Victoria is below England, they’re well below Singapore.

“So Victoria could be one of the top-performing systems in the world.”

Hunter said the key to lifting more children above the NAPLAN proficiency level was the ongoing training of their teachers.

“It’s not rocket science,” she said. “It comes down to investing in a much more strategic way in professional expertise for teachers, particularly in the primary school years. We should be doing more to lift the professional expertise of our primary teachers when it comes to mathematics, reading and writing.”

The Australian Education Union, which is embarking on what is expected to be a tough round of pay talks with the state government, said improved performance of Victoria’s children had been delivered by underpaid teachers and support staff.

“The premier and education minister are talking about these results, while cutting $2.4 billion in funding from Victoria’s public schools, and overseeing a system in which public school teachers work an average 12 unpaid overtime hours every week,” the union’s state president, Justin Mullaly, said.

At the conservative Centre for Independent Studies (CIS) think tank, education policy director Glenn Fahey said the results showed the nation’s education system was in a “state of high-spend stagnation”.

“Despite encouraging glimpses, the overwhelming evidence is of an education system stuck in neutral, when it needs to be revving up,” Fahey said. “The best that can be said for the education system is that Australia’s results haven’t dipped as badly as peer countries over the past few years, but that’s hardly something to celebrate.”

CIS research fellow Trisha Jha said the answer to mediocre performance was not more money, but a better use of time in the classroom.

“Governments have tried spending their way to school improvement, but that approach has reached its limits,” Jha said. “It’s time to shift the focus to what actually works: a knowledge-rich curriculum, explicit teaching, and better use of classroom time.”

Bethal Primary School in Melbourne’s north has overcome its serious socioeconomic challenges over the past three years, launching itself from the bottom 10 per cent of NAPLAN performers among comparable schools to be in the top 10 per cent this year.

Principal Dave Warren said the Meadow Heights school had been using explicit teaching and evidence-based learning long before it was mandated by the government, leading to standout results by the school’s year 3s and year 5s in reading, writing and numeracy.

“At our school, we don’t believe that postcode determines destiny,” Warren said on Tuesday. “Our staff are relentless in their belief that every child can achieve extraordinary things.

“When you combine that belief with passionate teaching and evidence-based practice, you can elevate student success beyond what many think is possible.”  Noel Towell is Education Editor for The Age

________________________________________________________________________________

National Assessment Program-Literacy and Numeracy (NAPLAN)).

Dr Kerry Hempenstall

My submission is in two parts addressing separate issues in the collection and use of data. The first relates to national assessment, and the second (provided as an attachment) is to describe individual and group assessments that have as a linking foundation a commitment to evidence-based practice in literacy data collection and instruction.

My intention in this first submission is to support national assessment. My focus is largely on literacy, as that has been the major domain of my work. In brief, my position is that national assessment is essential. The current format of NAPLAN may not be ideal, but it is certainly better than no national assessment at all. Self-serving and specious arguments should not be allowed to threaten this vital initiative. Worthwhile additions would include a national early identification scheme for Year One students, and providing greater transparency concerning the test raw scores that correspond to the various bands.

Some have argued that the literacy test is too limited, focussing as it does only on reading comprehension. While it might be helpful to assess other important components of reading, such as decoding and fluency, I see the main function of NAPLAN as acting as a national thermometer. It is to provide an indication as to the health of the patient – reading development across the nation’s students. Widening the test is both unnecessary and wasteful. It is unnecessary because when teachers are adequately retrained to use reading assessment properly, they can do the diagnostic testing on those students whose NAPLAN results are worrying. Making NAPLAN more extensive would take up additional student time, and would be a waste of time for all those students whose progress is quite adequate.

Before NAPLAN:

There is considerable interest and controversy in the Australian community concerning how our students are faring in the task of mastering reading, and how well our education system deals with teaching such a vital skill. Until state and subsequently national scale assessment commenced, we had largely small scale and disparate findings from researchers employing differing metrics to gauge how our students were progressing. Thus it was very difficult to make any judgements as to how students performed across the nation. Even the state based assessments that preceded NAPLAN were sufficiently different from each other to make comparisons fraught. Different emphases, formats, and scoring protocols were common. Defining where to establish the bands of competence varied significantly across studies. What does constitute an acceptable level of competence for this test or that test? Is it 50% correct - 80%, 25%?

Looking back one can see signs of this variation:

The Australian Government House of Representatives Enquiry (1993) estimated that between 10-20% of students finish primary school with literacy problems. In Victoria, about 16% were labelled reading disabled in two studies (Prior, Sanson, Smart, & Oberklaid, 1995; Richdale, Reece, & Lawson, 1996). Victorian Budget estimates (Public Accounts and Estimates Committee, 1999) anticipated that for the year 2000, 20% of Year 1 students would require the Reading Recovery program. Further concern was expressed that, after their Year Three at school, students with reading problems have little prospect of adequate progress (Australian Government House of Representatives Enquiry, 1993). Providing additional foundation for that fear was a Victorian study (Hill, 1995) that noted little discernible progress in literacy for the lowest 10% between Year Four and Year Ten. According to the Australian Council for Educational Research, more than 30% of Australian children entering high school (mainly in government and Catholic schools) could not read or write properly (“Our desperate schools”, 2000).

In more recent times, the Victorian Auditor General (2003) noted that Reading Recovery was being provided to on average 40-50% of Year 1 students, and in 2009 he reported that the budget for that program had climbed even higher each year since then. This gave pause for concern. How could so many students after a year or so of reading instruction require such extensive and costly remedial help? Could it be that initial reading instruction was not adequate? The introduction of state and then national assessment provided larger samples than were available during earlier times, and a troubling picture began to emerge.

National and international assessment

The Australian Bureau of Statistics in 2008 reported that almost half of all Australians aged 15-74 years had literacy skills below the level necessary to participate at even a basic level in our complex society. For example, they may have trouble managing transport timetable, completing basic forms, reading medicine labels, or interpreting weather maps. This indicates that any problems we have with literacy are not entirely new. How did we arrive at this position? One reason is that before broad-scale assessment our community was in the dark about the true state of affairs.

Reinforcing that alarming finding an OECD study (Programme for the International Assessment of Adult Competencies, PIAAC) revealed that:

Many adult Australians do not possess the literacy and numeracy skills necessary to participate fully in modern life and work. … Results from 2011-12 show that about 7.3 million or 44 per cent of adult Australians achieved in the lowest two bands for literacy, while about 8.9 million or 55 per cent achieved in the lowest two bands for numeracy (ACER, 2013).

There is clearly a concern that both national and international comparisons have not been flattering to us. There is a public perception that either educational outcomes for students have been further declining or that the education system is increasingly less able to meet rising community and employer expectations (Jones, 2012).

Parental concerns about literacy are becoming increasingly evident. In the Parents’ Attitudes to Schooling report (Department of Education, Science and Training, 2007), only 37.5% of the surveyed parents believed that students were leaving school with adequate skills in literacy. There has been an increase in dissatisfaction since the previous Parents' Attitudes to Schooling survey in 2003, when 61% of parents considered primary school education as good or very good, and 51% reported secondary education as good or very good. About 88 per cent of parents supported the use of a standard national process for assessing the skills of students. This level of NAPLAN support is markedly different to that expressed by many employed within the education field.

Press reports suggest that employers too have concerns about literacy development among young people generally, not simply for those usually considered to comprise an at-risk group (Australian Industry Group, 2016; Business Council of Australia, 2003; Collier, 2008).

This scrutiny of public education is not new; however, the focus in recent times has shifted as the real rate of student performance becomes common knowledge. Concerns that have arisen over recent years include apparent national and state test score declines, unflattering international achievement comparisons (ACER, 2010), the failure of large funding increases to produce discernible results (DEECD, 2012; Leigh & Ryan, 2008a; Nous Group Consortium, 2011), high school-dropout rates, and a perception that employment skill demands are not being adequately met by the education system (Collier, 2008; Levin, 1998). The press has displayed increased interest in highlighting these issues, thus further raising community awareness. Thus the expanding role for both national and international assessments has brought the issue further to public attention.

There appears now to be a public perception that there are serious problems in the education system’s capacity to meet their expectations. In the past, some teacher organisations have argued that the issues of student progress should be left in the hands of teachers, and that the school performance of students is quite acceptable compared with that of other nations. In my work - initially as a teacher, then as a school educational psychologist, and eventually in the RMIT Psychology Clinic over a forty year period, I experienced many examples of parents expressing dismay that their early concerns about their child’s progress were deflected by teachers. Usually, it was not until Year 4 or above that schools reported to parents that their child was struggling. This delay makes effective assistance far more difficult. One finding was that it takes four times as many resources to resolve a literacy problem by Year 4 than it does in Year 1 (Pfeiffer et al., 2001). This problem only intensifies as students approach upper primary and secondary school (Wanzek et al., 2013). Equally troubling was the Lesnick, Goerge, Smithgall, and Gwynne (2010) longitudinal study, which found that third grade students who were struggling with their reading had four times the rate of early school leaving compared with average readers.

In Australia, the broad scale assessment common for many years in the USA and GB is a more recent phenomenon. Begun initially at a State level, it has expanded to a national level - National Assessment Program Literacy and Numeracy (NAPLAN) and international level - the Progress in Reading Literacy Study (PIRLS) and the Programme for International Student Assessment (PISA).

What have recent analyses demonstrated?

  1. National

In the 2008 NAPLAN assessment, 19.6 per cent of Australian students were deemed at or below the national minimum standard in reading, and 18.7 per cent were at or below the standard in numeracy (Australian National Audit Office, 2012). This means that those below the national minimum standard require assistance, and so will some of those at the national minimum standard. How those national minimum standards are derived is unclear. The standards are supposed to be a snapshot of typical achievement. But it is not clear how typical achievement is defined. On what basis is a score of, say, 16/40 questions correct on a NAPLAN test considered typical achievement? Against which external criterion? This lack of information makes interpretation of results difficult. Another way of reporting the same results is to provide the figures for students who are at or above the national minimum standard. For 2012, the figure is 92% overall. This sounds much more satisfactory than around 19% at or below the standard, but how does it fit with the figures found in other studies that suggest, for reading, around 20% - 30% of students struggle? Something doesn’t compute.

Since 2008 there has been little change in average results – some national scores for some domains and for some Year levels have improved, while others have declined. Of course, there is also variation across states, and most dramatically, across socio-economic status and the indigenous/ non-indigenous divide (Australian Curriculum, Assessment and Reporting Authority, 2012.

  1. International

In 2012, the international PIRLS tests revealed that almost 25 per cent of Year 4 children in Australia failed to meet the standard in reading for their age. The report released by the Australian Council for Educational Research (ACER, 2012) reveal disappointing results for Australia in this latest international study of mathematics and science achievement, and in Australia’s first ever international assessment of reading at primary school level. Australian Year 4 students ranked 27th among the 53 nations involved, outperformed by other English-speaking nations such as England, the US and New Zealand. As this is the first time Australia has been involved in PIRLS, some consternation has followed the results.

Other international data (PISA) indicated a decline in reading (2000–2009) and mathematics (2003–2009) (Australian National Audit Office, 2012).

Although the OECD average for reading literacy has not changed between 2000 and 2009, ten countries have significantly improved their performance over this time, while five countries, including Australia, have declined significantly. … Australia’s reading literacy performance has declined, not only in terms of rankings among other participating countries but also in terms of average student performance. The mean scores for Australian students in PISA 2000 was 528 points, compared to 515 for PISA 2009. A decline in average scores was also noted between PISA 2000 and PISA 2006, when reading literacy was a minor domain (ACER, 2010).

Releasing the results, ACER Chief Executive Professor Geoff Masters said, “To say the results are disappointing is an understatement (ACER, 2012, p.1)”.

Learning to read written English.

The English written language is an amalgam of numerous other languages, such as Old English, Latin, Greek, German, French, and Old Norse. Because the different languages use differing letter combinations for the same spoken sound, reading written English is more difficult than in countries like Spain and Italy where one letter makes one sound (more or less). So, picking up reading unaided is a very difficult task. What then is a reasonable rate of reading difficulty? Is a 20-30% rate inevitable? Well, no. According to research, we should not be content until the reading difficulty rate falls to around 5% (Fuchs & Fuchs, 2005; Torgesen, 1998). Until then, we are not teaching reading well enough, and many students do not have an inbuilt resistance to learning how to read, but should be considered as instructional casualties.

This quote is from an interview with G. Reid-Lyon, Past-Chief of the Child Development and Behavior Branch of the National Institute of Child Health & Human Development, National Institutes of Health, USA.

When we look at the kids who are having a tough time learning to read and we went through the statistics, thirty-eight percent nationally, disaggregate that, seventy percent of kids from poverty and so forth hit the wall. Ninety-five of those kids are instructional casualties. About five to six percent of those kids have what we call dyslexia or learning disabilities in reading. Ninety-five percent of the kids hitting the wall in learning to read are what we call NBT: Never Been Taught. They’ve probably been with teachers where the heart was in the right place, they’ve been with teachers who wanted the best for the kids, but they have been with teachers who cannot answer the questions: 1) What goes into reading, what does it take? 2) Why do some kids have difficulty? 3) How can we identify kids early and prevent it? 4) How can we remediate it? (Boulton, 2003a).

This second quote is from an interview with Grover Whitehurst, Ex-Director (2002-08), Institute of Education Sciences, U.S. Department of Education, USA.

So, we have a difficult code, we have a neural system that for some children is not optimal for dealing with this code, and then we throw them an instructional system, a teaching system; teachers who don’t understand what the code really is or how it needs to be conveyed. And so the teacher is suggesting you should do this when in fact the child should be doing that. You can sample first or second grade classrooms around the country and you will still find, despite what we know about the process of reading and have learned over the past twenty years, you will still find that teachers for a first grader who is struggling to sound out a word who will discourage the child from doing that, and encourage the child to look at the pictures in the book and guess what that word means. Good readers don’t guess, good readers sound out almost every word on the page. And so the teacher is saying you solve the task this way when in fact the task has to be solved in an entirely different way. And that can not help but confuse children. So, non-optimal instruction, and in some cases simply misleading instruction, is a significant part of the problem (Boulton, 2003b).

Throwing money in the wrong direction

The Victorian Auditor General in 2009 reported that despite investing $1.19 billion in education initiatives over the previous six years there has been little improvement in average literacy and numeracy achievement. Leigh and Ryan (2011) showed that productivity (real expenditure vs student performance) decreased by 73 per cent between 1964 and 2003. This major component of this increase in expenditure has been in decreasing class sizes (a strategy shown elsewhere to be ineffective) (Jensen, 2012). It is now becoming clear that at a federal level learning must be shown to be a consequence of expenditure, and that requires a nationally based assessment system.

National and international assessments have the potential to provide some sense of how our children are faring in their education, or alternatively, how well our system teaches our children to read. However, there are limitations to the value of this style of testing, particularly when the only tasks included are those intended to assess reading comprehension. It is unquestionably a major function of reading, but not the only important component. The science of reading development has emphasised as additional areas: decoding, vocabulary, phonemic awareness, and reading fluency. To address this concern, one of the recommendations of the 2005 National Inquiry into the Teaching of Literacy (NITL) was that the NAPLAN “be refocused to make available diagnostic information on individual student performance, to assist teachers to plan for the most effective teaching strategies”. An example of a different focus for assessment is the Phonics Screening Check (Department of Education Schools, 2013) introduced into GB over the past couple of years (again against wide teacher complaint). It is held in mid-Year One and is designed to confirm whether students have learnt phonic decoding to an appropriate standard. As phonics is now considered an essential component of early reading instruction, this assessment is also an examination of how well schools are teaching reading, in addition to providing vital information to parents at this early stage. As a result 43 per cent of pilot schools were able to identify pupils with reading problems of which they were not already aware. This means that 235,000 pupils will now receive additional reading support from their schools, that might not have eventuated otherwise. Sub-standard results for a student are a red flag to schools to intervene early before the failure cascade called the Matthew Effect (Stanovich, 1986) becomes entrenched. Of course, when formal assessment does not commence until Year 3 (as in the NAPLAN), this opportunity in not available. Of course, this outcome does indicate how assessment alone does not have an impact on student progress. The will and the means to intervene efficiently and effectively must also accompany national assessment.

So, early assessment (assuming it leads to intervention) can have a prophylactic effect on reading failure, and is worthy of support. As to widening the scope of the NAPLAN reading assessment as the NITL recommended, I have a concern that it may make the assessment unwieldy. I agree entirely that diagnostic assessment is crucial for those students shown to be experiencing difficulty, but I am not sure that NAPLAN is the optimal vehicle for doing so.

Benchmark transparency

Because the benchmarks chosen for the various levels of proficiency in national assessments are not transparent, they are open to manipulation. Such occurrences have been reported in GB and the USA in the past (Adkins, Kingsbury, Dahlin, & Cronin, 2007). Also problematic is a slightly different issue - cheating by teachers and principals in the administration and scoring of tests (HuffPost Education, 2011; 2013). In recent times, similar claims of cheating and inappropriate test administration have been made in Australia. “Over the past three years, 164 schools around the country have sought to undermine the annual NAPLAN test by pressuring parents to withdraw their children, assisting students to complete the exam, or storing papers insecurely ahead of testing day” (Tomazin, 2013).

There can even be marked differences in results reported nationally and locally: “According to the National Assessment of Educational Progress (NAEP), the percentages of students who are proficient in basic reading and math are roughly half of the rates reported by the states” (Stone, 2013, p.5). This in itself makes a powerful case for national rather than solely state-based or locally-based assessments.

Over recent years in the USA, eight states had their reading and/or maths tests become significantly easier in at least two grades (Adkins, Kingsbury, Dahlin, & Cronin, 2007). The report, entitled The Proficiency Illusion, also found that recent improvements in proficiency rates on US state tests could be explained largely by declines in the difficulty of those tests.

So, a weakness of such opaque data is the potential for benchmarks to be manipulated to show governments of the day in the best possible light. There are examples in which benchmarks have been so low as to be at the level of chance. For example, when four multiple choice items constitute the response set for students, a 25% mark could be obtained by chance alone. Surely benchmarks would never be so low that chance alone could produce a proficiency level?

In 2006, the results needed to meet (Australian) national benchmarks for students in Years 3, 5 and 7 ranged from 22% to 44%, with an average of less than 34%. Year 3 students needed to achieve only 22% for reading, 39% for numeracy, and 30% for writing to be classified as meeting the minimum acceptable standard (Strutt, 2007, p.1).

Recently in Great Britain (Paton, 2008), the Assessment and Qualifications Alliance exam board admitted that standards had been lowered to elevate scores in 2008. In one exam paper, C grades (a good pass) were awarded to pupils who obtained a score of only 20%. Perhaps, having learned from this event, in the 2012 Phonics Screening Test, mentioned earlier, the pass mark was made public at 32 correct out of 40 questions, which is 80% correct.

Informal assessment

If community interest in literacy has been sparked, and there is public concern about the validity of the national broad scale assessment model, it is important for educators to offer guidance about high quality assessment. Part of the current literacy problem can be attributed to educators because they have not offered this high quality assessment in their schools to monitor progress. There has been a tendency to rely on informal assessment that often lacks validity and reliability (Watson, 1998), and unhelpful techniques like miscue analysis (Hempenstall, 1998), and the perusal of student folios (Fehring, 2001).

In a three-year Australian study:

Wyatt-Smith and Castleton investigated how Australian teachers made judgments about student writing using literacy benchmark standards (Department of Employment, Education, Training and Youth Affairs [DEETYA] 1998; Wyatt-Smith and Castleton 2005). … Teachers made judgments based on their own experience; explicit literacy standards were not part of teachers' experience, and teachers accepted that their "in head" standards varied from year to year and from class to class” (Bolt, 2011, p.158).

Studies by Feinberg and Shapiro (2009) and by Bates and Nettlebeck (2001) each reported that informal assessments were significantly less accurate for struggling readers when compared with formal assessment, and most teachers’ over-estimation among students with low achievement scores was greater than a year of reading age. In the Madelaine and Wheldall (2003) study, Australian teachers not only failed to identify 10% of the struggling students, but 18% of them also falsely reported as struggling some students who were not doing so. Limbos and Geva (2001) found that teachers tended to ascribe adequate reading skills to struggling readers on the basis of their reasonable oral language skills. That is, without adequate reading assessment, teachers can be fooled into incorrectly assuming that a student’s literacy must be OK because they express their thoughts well in speech.

The Productivity Commission in 2012 expressed concern that:

Because teachers are kept in ignorance of specific learning difficulties, students are under diagnosed and under supported. Teachers are not able to recognise the signs which should lead to testing by a psychologist or specialist in specific learning difficulties. Furthermore, they often don’t know who the student should be referred to. (sub. DR76, p. 2) (Productivity Commission, 2012, p.284).

If every teacher did implement a standard, agreed upon assessment schedule, based upon the current evidence on reading development, then one might argue against the need for national assessment. Data could be collected from schools, and would be comparable across the nation - based upon a similar metric. Of course, the problem of variation in teachers’ test presentation skills would remain. An important aspect of the current system is to provide consistency for the more than 100,000 students who change schools across State/Territory boundaries, sectors and regions (National Inquiry into the Teaching of Literacy, 2005).

The assessment of important literacy components can supply valuable information not available in the current broad scale testing program. For example, assessment can assist in the identification and management of students at-risk even before reading instruction commences. They can also help identify those making slow progress at any year level. This is especially important given the usually stable learning trajectory from the very early stages. If specific interventions are implemented, appropriate reading assessment can provide on-going information about the effectiveness of the chosen approach. There is an important question implicit in this potentially valuable activity. What sorts of assessment are likely to be most beneficial in precluding reading pitfalls and enhancing reading success? In this submission, the emphasis is directed towards assessment of those aspects of reading that have been identified by research as critical to reading development.

From assessment to intervention

It is recognised that literacy assessment itself has little intrinsic value; rather, it is only the consequences flowing from the assessment process that have the potential to enhance the prospects of those students currently struggling to master reading. Assessment also allows for the monitoring of progress during an intervention, and evaluation of success at the end of the intervention. However, the initial value relates to the question of whether there is a problem, and if so, what should be done. What should be done is inevitably tied to the conception of the reading process, and what can impede its progress. How do educationists tend to view the genesis of reading problems?

Perceptions of literacy problems and causes

Alessi (1988) contacted 50 school psychologists who, between them, produced about 5000 assessment reports in a year. The school psychologists agreed that a lack of academic or behavioural progress could be attributed to one or more of the five factors below. Alessi then examined the reports to see what factors had been assigned as the causes of their students’ educational problems.

1. Curriculum factors? No reports.

2. Inappropriate teaching practices? No reports.

3. School administrative factors? No reports.

4. Parent and home factors? 10-20% of reports.

5. Factors associated with the child? 100%.

In another study this time surveying classroom teachers, Wade and Moore (1993) noted that when students failed to learn 65% of teachers considered that student characteristics were responsible while a further 32% emphasised home factors. Only the remaining 3% believed that the education system was the most important factor in student achievement, a finding utterly at odds with the research into teacher effects both in Australia and overseas (Cuttance, 1998; Hattie, 2009; Hattie, Clinton, Thompson, & Schmidt-Davies, 1995).

This highlights one of the ways in which assessment can be unnecessarily limiting in its breadth, if the causes of students’ difficulties are presumed to reside solely within the students, rather than within the instructional system. Assessment of students is not a productive use of time unless it is carefully integrated into a plan involving instructional action.

When the incidence of failure is unacceptably high, as in the USA, GB, and Australia, then an appropriate direction for resource allocation is towards the assessment of instruction. It can only be flawed instruction that intensifies the reading problem from a realistic incidence of reading disability of around 5% (Brown & Felton, 1990; Felton, 1993; Marshall & Hynd, 1993; Torgesen, Wagner, Rashotte, Alexander, & Conway, 1997; Vellutino et al., 1996) to that which we find of 20 - 30% (or higher). A tendency can develop for victim blame. "Learning disabilities have become a sociological sponge to wipe up the spills of general education. … It's where children who weren't taught well go (p.A1)" (Lyon, 1999).

Evidence-based assessment and practice

There is an increasing recognition that an education system must constantly assess the quality of instruction provided in its schools, and that it should take account of the findings of research in establishing its benchmarks and policies. “Thus the central problem for a scientific approach to the matter is not to find out what is wrong with the children, but what can be done to improve the educational system” (Labov, 2003, p.128). The development of an Australian national English curriculum is an example of this emerging system interest. Up to this time, education systems in Australia have been relatively impervious to such findings (Hempenstall, 1996, 2006), lagging behind significant (if tenuous) changes in the USA with Reading First (Al Otaiba et al., 2008) and in Great Britain with the Primary National Strategy (2006).

Unfortunately, Australia has been lax in evaluating the effectiveness of the instructional approaches it has supported (Rorris et al., 2011).

Re students from disadvantaged groups, learning disabilities, indigenous, ESL, low SES, remote areas. Weak monitoring and reporting inhibits the capacity of school systems to build sector knowledge of the relevance and context of improvement strategies that have demonstrated effectiveness. This means there is a lack of evidence-based links for programs and their effects on learning (254). …  There are insufficient data available to establish to what extent existing programs are effective because few have been evaluated, and fewer still have been evaluated with student outcomes as a focus” (p.87).

The Victorian Auditor-General (2009, 2012) found that there “has been no system-wide assessment of the ongoing effectiveness of key elements of the approach, such as the Reading Recovery intervention” (p.5) …. Further, “DEECD does not consistently use monitoring, program reviews and evaluations” (p.57).

Clearly, the education system failed to properly evaluate the impact of the programs it has implemented. One proposed reason offered by the Productivity Commission (2012)

 … in order to ‘save face’ policymakers may continue with programs that are demonstrably poor investments rather than abandoning underperforming policies because - in acknowledging the results of any evaluation - they might be accused of ‘failure’”. … However, changing policies without the benefit of evidence offers no assurance that outcomes will improve. In fact, a cycle of constantly changing policies can be potentially destructive where it fosters instability and reduces confidence (particularly within disadvantaged communities) in the education system. Evaluation is the first step towards greater continuity in the policy and institutional landscape — a necessary, although not sufficient, condition for achieving sustained advances in education outcomes. As such, it is essential that policymakers subject their initiatives to proper analysis, and do not move on to another policy idea before evaluations can be conducted (p.299).

One means of encouraging greater scrutiny is being able to examine the impact programs have on student performance at the national level.

Even allowing that the major problem for the education system lies in the realm of instruction, particularly in the initial teaching of reading, individual student assessment remains of value. It is, of course, necessary as a means of evaluating instructional adequacy. Beyond that, there is great potential value in the early identification of potential reading problems, in determining the appropriate focus for instruction, in the monitoring of progress in relevant skill areas, and with the evaluation of reading interventions. It is the assumption in this paper that decisions about assessment should be driven by up-to-date conceptions of the important elements in reading development.

The extent of the literacy problem has been known to literacy researchers for many years, and a group of 26 Australian researchers prompted the 2005 National Inquiry into the Teaching of Literacy (NITL). The findings included 20 recommendations for reform – none of which was adopted by the government of the day or since.

It is more difficult for governments and education departments to ignore the results of national and international assessments; thus, we are beginning to see strong words and (perhaps) action, but how and to what effect become crucial questions.

In 2012, a larger group of researchers (with including many of the original 2004 group) responded to recent findings of national and international assessment with a reminder of what is known about the causes of the literacy problem:

In an open letter to federal, state and territory education ministers and their opposition counterparts, a group of 36 educators, scientists and clinicians call for a ``vast shake-up at all levels of teacher training'' to ensure children are taught to read properly. The letter was prompted by the results in the international Progress in Reading Literacy Study tests last week that revealed almost 25 per cent of Year 4 children in Australia failed to meet the standard in reading for their age, to the shock of many educators and governments. Reprising the letter they sent to then education minister Brendan Nelson in 2004 that resulted in the independent inquiry, the researchers admonish governments for their collective failure to heed the evidence and advice for almost a decade on the most effective way to teach reading (Ferrari, 2012).

A segment of that letter is reprinted below:

We have significant problems in education from the beginning stages, in that we do not teach reading well. We do not use approaches known to be effective in initial reading instruction. As a nation, we do not routinely screen students entering school for underdeveloped pre-reading skills critical for facilitating learning to read, nor do we monitor student progress in learning to read in a manner that allows for successful early intervention with students failing to progress. We do not redress our early system failure during the middle primary years. In the secondary years, we have a significant group of disillusioned students who have lost contact with the curriculum because of these earlier issues. We tend to focus attention and resources upon compensatory educational options instead of emphasising the resolution of our earlier mistakes. The sequence of initial failure-shame-frustration-disengagement-dropout is predictable and ongoing. Currently, it is being addressed piecemeal, as if they were separate problems.

We need a vast shake-up at all levels of teacher training. By turning our gaze to educational practices supported by empirical research we can make optimum use of our resources to complete the task with effectiveness and efficiency (Coltheart, Hempenstall, Wheldall, et al., 2012).

Inadequate teacher training of evidence-based assessment and instruction.

There is an increasing recognition that teacher training has produced a teaching force that has not been well prepared to provide effective teaching to our diverse student community (Malatesha Joshi et al., 2009). Why this is so is a long story, but a history is available at http://www.educationnews.org/articles/the-whole-language-phonics-controversy-a-historical-perspective.html

Evidence for this assertion follows:

All too often Victoria’s teacher training, referred to as pre-service education, falls short of the demands of today’s schools. While there are many providers, quality outcomes are inconsistent. Principals report that in the case of more than one-third of teachers, insufficient pedagogical preparation hinders student instruction (p.10). … At present less than 30 per cent of principals feel new teachers are well prepared to communicate with parents, manage classroom activities well, and provide effective support and feedback to students, which are all largely recognised as important skills for effective teaching and learning.    Only around half of graduates report satisfaction with the preparation provided by their courses (DEECD, 2012, p.11).

A survey by Rohl and Greaves (2005) reported that 36% of beginning primary teachers felt unprepared to teach reading and 57% unprepared to teach phonics. Senior staff at their schools were more pessimistic, considering that 49% of these beginning teachers were unprepared to teach reading, and 65% unprepared to teach phonics. These figures on unpreparedness rose dramatically (77% - 89%) when the beginning teachers were confronted with diverse learners (those with disabilities or learning difficulties, indigenous and low SES students, and students whose initial language was not English). Other Australian studies by Maher and Richdale (2008) and by Fielding-Barnsley (2010) noted similar results.

Several other Australian studies also support these findings:

Taken together, these results indicate that for this cohort of pre-service teachers, entry knowledge of graphological/phonological rules and terminology tends to be fragmentary, suggesting that without further instruction in domain-specific knowledge in the area of phonological awareness and phonics, they may have difficulty providing systematic and explicit beginning reading instruction. This supports findings from previous studies which found that many pre-service and in-service teachers have limited knowledge of phonological awareness and phonics (e.g. Fielding-Barnsley & Purdue, 2005; Moats & Foorman, 2003; Rennie & Harper, 2006; Rohl & Greaves, 2005). … The written comments have also highlighted, unintentionally, the fact that a number of the pre-service teachers in the present study, like those in reported by Zipin and Brennan (2006), showed deficiencies in personal literacy skills with regard to grammar, punctuation, and sentence structure (Fisher, Bruce, & Greive, 2007, p.82-3, 85).

This situation was recognized in the 2005 recommendations of the National Inquiry into the Teaching of Literacy: “1. The Committee recommends that teachers be equipped with teaching strategies based on findings from rigorous, evidence-based research that are shown to be effective in enhancing the literacy development of all children”. Unfortunately, as with all the other recommendations of this important report, it was ignored, and the report removed from the Australian Government website. It can still be read at the website of the Australian Council for Educational Research – see http://tinyurl.com/d6v2v9y

One reason proposed for this situation is the failure of education faculties to take the teaching of reading seriously in their course planning.

Responses to the national survey indicate that in almost all of the nominated courses, less than 10 per cent of time in compulsory subjects/units is devoted to preparing student teachers to teach reading. They also indicated that in half of all courses less than five per cent of total instructional time is devoted to teaching reading (National Inquiry into the Teaching of Literacy, p.37).

Teacher quality

Creating further concern is the trend towards accepting into teacher training students whose aptitude is substantially lower than in the past.

We find that the aptitude of new teachers has fallen considerably. Between 1983 and 2003, the average percentile rank of those entering teacher education fell from 74 to 61, while the average rank of new teachers fell from 70 to 62. We find that two factors account for much of the decline: a fall in average teacher pay (relative to other occupations) and a rise in pay differentials in non-teaching occupations (Leigh & Ryan, 2008b, p.141).

Further, “Some of the biggest teaching schools are accepting entry-level students with TER scores so low as to be equivalent to failure in other states” (Senate Employment, Workplace Relations and Education Committee, 2007, p.7).

One consequence has been an apparent decline in student-teacher literacy to the extent that their capacity to teach in an evidence-based manner has been questioned.

The literacy competency of student teachers was raised as an issue in all focus group discussions. Respondents reported that many students lacked the literacy skills required to be effective teachers of reading. These students needed help to develop their foundational literacy skills. They also needed explicit teaching about meta-linguistic concepts, for example, phonemic awareness, phonics, and the alphabetic principle (NITL, p.33).

This led to another recommendation:

Recommendation 14

The Committee recommends that the conditions for teacher registration of all primary and secondary graduates include a demonstrated command of personal literacy skills necessary for effective teaching, and a demonstrated ability to teach literacy within the framework of their employment/teaching program (p.35).

So, there is a problem with teacher training that, unless overcome, will hinder efforts to raise the standards of school education. Stronger accreditation guidelines for training institutions have been requested for a number of years. There is some evidence of that occurring, and also recently attention is beginning to be directed at the entry criteria into teacher training courses.

Resistance to NAPLAN

A problem threatening the NAPLAN at present is the resistance from a number of teachers and teacher organisations either to national assessment per se, or to its current format. Some of this resistance has led to unprofessional behaviour, such as attempting to “teach to the test” and worse. The Australian Curriculum, Assessment and Reporting Authority (ACARA) noted cheating and other substantiated breaches of general administration protocols during NAPLAN tests over several years and in all states. These events, though modest considering the breadth of the exercise, have increased from 63 in 2010 to 74 in 2011, and to 98 in 2012 (Australian Curriculum, Assessment and Reporting Authority, 2013). The increases are of concern because they parallel events in both GB and the USA that have become much more widespread.

Last year an investigation into the largest ever cheating scandal in US schools revealed rampant, systematic cheating on test scores in Atlanta’s public schools. It found that 178 teachers and principals in 40 of Atlanta’s 100 public schools cheated on state standardized tests in 2009. It uncovered instances of cheating dating back to 2001. The report said that extreme pressure to boost test scores drove teachers and principals to cheat. An investigation is ongoing into high erasure rates on mathematics and reading tests in more than 100 schools in Washington DC. In England last year, even exam boards were caught cheating by providing test questions to teachers months before the exams were due. In 2009, students from 70 schools had their Sats test results annulled or changed because of cheating by teachers or bungled handling of the exams. Reported cheating incidents are notoriously under-estimates of the real situation. “It’s just the tip of the iceberg, I think,” says US testing expert Professor Tom Haladyna, “The other 80 percent is being hidden” [Atlanta Journal-Constitution, 21 June 2009] (Cobbold, 2012).

In the USA, the biennial National Assessment of Educational Progress (NAEP) is considered the gold standard, first because it is divorced from the decisions about student retention or teacher/principal job security that are consequences of state assessments. Second, the procedures used by the employed independent contractors who oversee the administration and transportation of the assessments make it practically impossible for teachers and principals to change the students’ work on these tests.

I do not intend to counter the many calls for an end to NAPLAN, as I believe they arise out of ignorance, self-serving, or fear. An excellent rebuttal of many of the criticisms (Righting some wrong thinking on NAPLAN) was published recently in The Age by Philip Henseleit, and can be found at http://www.theage.com.au/comment/righting-some-wrong-thinking-on-naplan-20130515-2jmns.html

We can and must teach reading more effectively. So, few teachers have been trained in the evidence-based reading instruction espoused in almost every significant report over the past decade (National Early Literacy Panel, 2008; National Inquiry into the Teaching of Literacy, 2005; National Reading Panel, 2000, Primary National Strategy, 2006; Rose Report, 2006). Few teachers know how to assess reading in a scientific manner, or know how and why to read and interpret research and data. Today, in an era of professional accountability, teaching remains a craft rather than a science, a guild rather than a profession. Many teachers have been inculcated with a constructivist philosophy asserting that students will necessarily interpret instruction in their own unique ways. This has arisen in their training, and teachers have been diverted from effective literacy instruction by the rhetoric about constructivism, multi-literacies, inclusion, differentiated instruction, personalised learning, critical literacy, brain-based learning, discovery learning, whole language, and learning styles. To overcome these obstacles to the effective teaching of reading will not be easy, but one pre-condition is that we have regular data nationally about the progress of our teaching efforts.

It’s hardly a revelation to argue that the adoption of evidence-based practice (EBP) in some other professions is far advanced in comparison to its use in education. That’s not to say that the resistance displayed by some teacher organizations towards the adoption of EBP has not been evident in the early stages of its acceptance by those professions, such as medicine and psychology. However, as these principles have been espoused in medicine and psychology since the early nineties, a new generation of practitioners have been exposed to EBP as the normal standard for practice. This has occurred among young practitioners because their training has emphasized the centrality of evidence in competent practice.

In education, unfortunately, there are few signs of this sequence occurring. Most teachers-in-training are not exposed to either the principles of EBP (unless in a dismissive aside) or to the practices that have been shown to be beneficial to student learning, such as the principles of instructional design and effective teaching, explicit phonological instruction, and student management approaches that might be loosely grouped under a cognitive-behavioural banner.

In my view, until educational practice includes EBP as a major determinant of practice, then it will continue to be viewed as an immature profession. It is likely that the low status of teachers in many western countries will continue to be the norm unless and until significant change occurs. A fundamental component of professions is the systematic collection of data to inform decision-making. The acceptance by the teaching profession of this tenet of EBP may be promoted by the gradual acceptance of national assessment.

I conclude by re-affirming my support for national assessment, but with modifications consistent with the recommendations of the National Inquiry into the Teaching of Literacy (2005). These include:

9. The Committee recommends that the teaching of literacy throughout schooling be informed by comprehensive, diagnostic and developmentally appropriate assessments of every child, mapped on common scales. Further, it is recommended that: • nationally consistent assessments on-entry to school be undertaken for every child, including regular monitoring of decoding skills and word reading accuracy using objective testing of specific skills, and that these link to future assessments, including the use of regular monitoring of decoding skills and word reading accuracy using objective testing of specific skills;• • education authorities and schools be responsible for the measurement of individual progress in literacy by regularly monitoring the development of each child and reporting progress twice each year for the first three years of schooling.

References:

ACER (2010). PISA in Brief. Highlights from the full Australian Report: Challenges for Australian Education: Results from PISA 2009. Retrieved from http://www.acer.edu.au/documents/PISA-2009-In-Brief.pdf

ACER (2012). ACER releases results from latest international studies of student achievement. Australian Council for Educational Research. Retrieved from http://www.acer.edu.au/media/acer-releases-results-from-latest-international-studies-of-student-achievem

ACER (2013). International study reveals serious adult literacy and numeracy problems. Retrieved from http://www.acer.edu.au/media/international-study-reveals-serious-literacy-and-numeracy-problems

Adkins, D., Kingsbury, G.G., Dahlin, M., & Cronin, J. (2007). The proficiency illusion. Thomas B. Fordham Institute. Retrieved from http://www.edexcellence.net/publications/theproficiencyillusion.html

Al Otaiba, S., Connor, C., Lane, H., Kosanovich, M. L., Schatschneider, C., Dyrlund, A. K., Miller, M. S., & Wright, T. L. (2008). Reading First kindergarten classroom instruction and students' growth in phonological awareness and letter naming–decoding fluency. Journal of School Psychology, 46(3), 281-314.

Alessi, G. (1988). Diagnosis diagnosed: A systemic reaction. Professional School Psychology, 3, 145-151.

Australian Bureau of Statistics. (2008). Adult Literacy and Life Skills Survey (ALLS 2006-2007). Retrieved from http://www.abs.gov.au/AUSSTATS/abs@.nsf/Lookup/4102.0Chapter6102008

Australian Curriculum, Assessment and Reporting Authority (ACARA) (2013). Report of 2012 NAPLAN test incidents. Retrieved from http://www.acara.edu.au/verve/_resources/2012_NAPLAN_TEST_INCIDENTS_REPORT_website_version.pdf#search=cheating

Australian Curriculum, Assessment and Reporting Authority (ACARA) (2012). National Assessment Program Literacy and Numeracy NAPLAN Summary Report. Preliminary results for achievement in Reading, Writing, Language Conventions and Numeracy. Retrieved from http://www.acara.edu.au/default.asp

Australian Government House of Representatives Enquiry. (1993). The literacy challenge. Canberra: Australian Printing Office.

Australian Industry Group. (2016). Tackling foundation skills in the workforce. Ai Group, Sydney. Retrieved from http://www.aigroup.com.au/portal/binary/com.epicentric.contentmanagement.servlet.ContentDeliveryServlet/LIVE_CONTENT/Publications/Reports/2016/AIG9675_EMAIL.pdf

Australian National Audit Office. (2012). National Partnership Agreement on Literacy and Numeracy.

Bates, C., & Nettlebeck, T. (2001). Primary school teachers’ judgements of reading achievement. Educational Psychology, 21(2), 179-189.

Bolt, S. (2011). Making consistent judgments: Assessing student attainment of systemic achievement targets. The Educational Forum, 75(2), 157-172. Retrieved from http://search.proquest.com/docview/863245835?accountid=13552

Boulton, D. (2003a). Children of the Code Interview with G. Reid-Lyon, Past-Chief of the Child Development and Behavior Branch of the National Institute of Child Health & Human Development, National Institutes of Health. Retrieved from http://www.childrenofthecode.org/interviews/lyon.htm#Instructionalcasualties

Boulton, D. (2003b). Children of the Code Interview with Grover Whitehurst, Ex-Director (2002-08), Institute of Education Sciences, U.S. Department of Education. Retrieved from http://www.childrenofthecode.org/interviews/whitehurst.htm#WhyisReadingsoDifficult

Brown, I. S. & Felton, R. H. (1990). Effects of instruction on beginning reading skills in children at risk for reading disability. Reading & Writing: An Interdisciplinary Journal, 2, 223-241.

Business Council of Australia. (2003). The cost of dropping out: The economic impact of early school leaving. Retrieved from http://www.bca.com.au/upload/The_Cost_of_Dropping_Out.pdf

Cobbold, T. (2012). Fighting for equity in education: Several schools found to be cheating in NAPLAN Tests. Save Our Schools. Wednesday January 18, 2012. Retrieved from http://www.saveourschools.com.au/league-tables/several-schools-found-to-be-cheating-in-naplan-tests

Collier, K. (2008, October 18). The ABC of ignorance. Herald Sun, p.9.

Coltheart, M., Hempenstall, K., Wheldall, K., et al. (2012). An open letter to all Federal and State Ministers of Education. Retrieved from http://tinyurl.com/dytdcyu

Cuttance, P. (1998). Quality assurance reviews as a catalyst for school improvement in Australia. In A. Hargreaves, A. Lieberman, M. Fullan., & D. Hopkins (Eds.), International handbook of educational change, Part II (pp. 1135-1162). Dordrecht: Kluwer Publishers.

DEECD. (2012). New directions for school leadership and the teaching profession: Discussion paper, June 2012

Department of Education Schools. (2013). Phonics screening check materials. Retrieved from http://www.education.gov.uk/schools/teachingandlearning/assessment/keystage1/a00200415/phonics

Department of Education, Science and Training. (2007). Parents’ attitudes to schooling. Canberra: Australian Government.http://www.dest.gov.au/NR/rdonlyres/311AA3E6-412E-4FA4-AC01-541F37070529/16736/ParentsAttitudestoSchoolingreporMay073.rtf

Fehring, H. (2001). Literacy assessment and reporting: Changing practices. 12th European Conference on Reading, RAI Dublin, 1st - 4th July. Retrieved from http://sece.eu.rmit.edu.au/staff/fehring/irish.htm

Feinberg, A. B., & Shapiro, E. S. (2009). Teacher accuracy: An examination of teacher-based judgments of students reading with differing achievement levels. The Journal of Educational Research, 102(6), 453-462, 480.

Felton, R. H. (1993). Effects of instruction on the decoding skills of children with phonological-processing problems. Journal of Learning Disabilities, 26, 583-589.

Ferrari, J. (2012). A decade of lost action on literacy. The Australian, 22 December 2012, p.1.

Fielding-Barnsley, R. (2010): Australian pre-service teachers' knowledge of phonemic awareness and phonics in the process of learning to read. Australian Journal of Learning Difficulties, 15(1), 99-110

Fisher, B.J., Bruce, M.E., & Greive, C. (2007). The entry knowledge of Australian pre-service teachers in the area of phonological awareness and phonics. In A Simpson (Ed.), Future directions in literacy: International conversations 2007. University of Sydney. Retrieved from http://ses.library.usyd.edu.au/bitstream/2123/2330/1/FutureDirections_Ch5.pdf

Fuchs, D., & Fuchs, L.S. (2005). Peer-assisted learning strategies: Promoting word recognition, fluency, and reading comprehension in young children. The Journal of Special Education 39(1), 34-44.

Hattie J.A. (2009). Visible learning: A synthesis of over 800 meta-analyses relating to achievement. London, UK: Routledge.

Hattie, J. A., Clinton, J., Thompson, M., & Schmidt-Davies, H. (1995). Identifying highly accomplished teachers: A validation study. Greensboro, NC: Center for Educational Research and Evaluation, University of North Carolina.

Hempenstall, K. (1996). The gulf between educational research and policy: The example of direct instruction and whole language. Behaviour Change, 13, 33-46.

Hempenstall, K. (1998). Miscue analysis: A critique. Australian Journal of Learning Disabilities, 3(4), 32-37.

Hempenstall, K. (2006). What does evidence-based practice in education mean? Australian Journal of Learning Disabilities, 11(2), 83-92.

Hill, P. (1995). School effectiveness and improvement: Present realities and future possibilities. Dean's Lecture: Paper presented at Melbourne University, May 24. Retrieved from http://www.edfac.unimelb.edu.au/Seminars/dean_lec/list.html

HuffPost Education. (2013, April 22). Beverly Hall's lawyers deny the schools chief had role in Atlanta cheating scandal. Retrieved from http://www.huffingtonpost.com/2013/04/22/atlanta-cheating-scandal-beverly-hall_n_3132583.html

Jensen, B. (2012). Targeting the things that matter. Grattan Institute, Vic. ACER Conference “School Improvement: What does research tell us about effective strategies?” Retrieved from http://research.acer.edu.au/research_conference/RC2012/28august/8

Jones, J.M. (2012). Confidence in U.S. public schools at new low. Gallup Politics. Retrieved from http://www.gallup.com/poll/155258/Confidence-Public-Schools-New-Low.aspx

Labov, L. (2003). When ordinary children fail to read. Reading Research Quarterly, 38, 128-131.

Leigh, A & Ryan, C. (2008b). How and why has teacher quality changed in Australia? Australian Economic Review, 41(2), 141-159. Retrieved from http://digitalcollections.anu.edu.au/handle/1885/45254

Leigh, A., & Ryan, C. (2008a). How has school productivity changed in Australia? . The Australian National University, Canberra. Retrieved from http://econrsss.anu.edu.au/~aleigh/pdf/SchoolProductivity.pdf

Leigh, A., & Ryan, C. (2011). Long-run trends in school productivity: Evidence from Australia. Education Finance and Policy, 6(1), 105–135.

Lesnick, J., Goerge, R., Smithgall, C. & Gwynne, J. (2010). Reading on grade level in third grade: How is it related to high school performance and college enrollment? Chapin Hall; Consortium on Chicago School Research; The Annie E. Casey Foundation. Retrieved from http://www.aecf.org/KnowledgeCenter/Publications.aspx?pubguid={61221250-BC02-49C9-8BDA-D64C45B1C80C}

Levin, B. (1998). Criticizing the schools: Then and now. Education Policy Analysis Archives, 6(16). Retrieved from http://epaa.asu.edu/epaa/v6n16.html

Limbos, M., & Geva, E. (2001). Accuracy of teacher assessments of second-language students at risk for reading disability Journal of Learning Disabilities, 34, 136-151.

Lyon, G. R. (1999, December, 12). Special education in state is failing on many fronts. Los Angeles Times, p. A1. Retrieved from http://www.latimes.com/news/state/reports/specialeduc/lat_special991212.htm

Madelaine, A., & Wheldall, K. (2003). Can teachers discriminate low-progress readers from average readers in regular classes? Australian Journal of Learning Disabilities, 8(3), 4-7.

Maher, N., & Richdale , A. (2008) Primary teachers’ linguistic knowledge and perceptions of early literacy instruction. Australian Journal of Learning Disabilities, 13(1), 17-37.

Malatesha Joshi, R., Binks, E., Hougen, M., Dahlgren, M. E., Ocker-Dean, E., & Smith, D. L. (2009). Why elementary teachers might be inadequately prepared to teach reading. Journal of Learning Disabilities, 42(5), 392-402

Marshall R. M., & Hynd, G. W. (1993). Neurological basis of learning disabilities. In William W. Bender (Ed.) Learning disabilities: Best practices for professionals. Stoneham, MA: Butterworth-Heinemann.

National Early Literacy Panel. (2008). Developing Early literacy: Report of the National Early Literacy Panel. Washington DC: National Institute of Literacy. Retrieved from http://www.nifl.gov/earlychildhood/NELP/NELPreport.html

National Inquiry into the Teaching of Literacy. (2005). Teaching Reading: National Inquiry into the Teaching of Literacy. Canberra: Department of Education, Science, and Training. Retrieved from http://tinyurl.com/d6v2v9y

National Reading Panel. (2000). National Reading Panel: Teaching children to read. Retrieved from http://www.nationalreadingpanel.org.

Nous Group Consortium. (2011). Schooling challenges and opportunities: A report for the Review of Funding for Schooling Panel. Nous Group Consortium, August 29. http://foi.deewr.gov.au/documents/schooling-challenges-and-opportunities

Office of the Victorian Auditor General. (2003). Improving literacy standards in government schools. Retrieved from http://www.audit.vic.gov.au/reports_par/Literacy_Report.pdf

Office of the Victorian Auditor-General (2012). Programs for students with special learning needs: Audit summary. Retrieved from http://www.audit.vic.gov.au/publications/20120829-Special-Learning-Need/20120829-Special-Learning-Need.rtf

Office of the Victorian Auditor-General. (2009). Literacy and numeracy achievement. Retrieved from http://www.audit.vic.gov.au/reports__publications/reports_by_year/2009/20090204_literacy_numeracy/1_executive_summary.aspx

Our desperate school. (2000, August 8). The Age, p.11.

Paton, G. (2008, 25 Oct). GCSE standards 'deliberately lowered' to make sure pupils pass. Telegraph.co.uk. Retrieved from http://www.telegraph.co.uk/news/newstopics/politics/education/3254233/GCSE-standards-deliberately-lowered-to-make-sure-pupils-pass.html

Pfeiffer, S., Davis, R., Kellog, E., Hern, C., McLaughlin, T.F., & Curry, G. (2001).The effect of the Davis Learning Strategies on First Grade word recognition and subsequent special education referrals. Reading Improvement, 38(2), 1-19.

Primary National Strategy (2006). Primary framework for literacy and mathematics. UK: Department of Education and Skills. Retrieved from http://www.standards.dfes.gov.uk/primaryframeworks/

Prior, M., Sanson, A. Smart, D., & Oberklaid, F. (1995). Reading disability in an Australian community sample. Australian Journal of Psychology, 47(1), 32-37.

Productivity Commission. (2012). Schools Workforce, Research Report, Canberra.  JEL code: I21, I28, J24. Retrieved from http://www.pc.gov.au/projects/study/education-workforce/schools/report

Public Accounts and Estimates Committee (1999). Report on the 1999-2000 Victorian Budget Estimates. Retrieved from http://www.parliament.vic.gov.au/paec/33Report.pdf

Richdale, A. L., Reece, J. E., & Lawson, A. (1996). Teachers, children with reading difficulties, and remedial reading assistance in primary schools. Behaviour Change, 13(1), 47-61.

Rohl, M., & Greaves, D. (2005). How are pre-service teachers in Australia being prepared for teaching literacy and numeracy to a diverse range of students? Australian Journal of Learning Disabilities, 10(1), 3-8.

Rorris, A., Weldon, P., Beavis, A., McKenzie, P., Bramich, M., & Deery, A. (2011). Assessment of current process for targeting of schools funding to disadvantaged students. An Australian Council for Educational Research report prepared for The Review of Funding for Schooling Panel. Retrieved from http://www.deewr.gov.au/Schooling/ReviewofFunding/Pages/PaperCommissionedResearch.aspx

Rose, J. (2006). Independent review of the teaching of early reading. Bristol: Department for Education and Skills. Retrieved from www.standards.dfes.gov.uk/rosereview/report.pdf

Senate Employment, Workplace Relations and Education Committee. (2007). Quality of school education. Commonwealth of Australia. Retrieved from http://www.aph.gov.au/SEnate/committee/eet_ctte/completed_inquiries/2004-07/academic_standards/index.htm

Stanovich, K. E. (1986). Matthew effects in reading: Some consequences of individual differences in the acquisition of literacy. Reading Research Quarterly, 21, 360-406.

Stone, J.E. (2013). Reversing American decline by reducing education’s casualties: First, we need to recapture our school boards. Education Consumers Foundation. Retrieved from http://www.education-consumers.org/rad.htm

Strutt, J. (2007). Students fail on 'three Rs' test. The West Australian, Monday 10 December. Retrieved from http://www.platowa.com/Breaking_News/2007/2007_12_10.html

Tomazin. F. (2013, Feb 17). Schools caught cheating on NAPLAN. The Age. Retrieved from http://www.theage.com.au/victoria/schools-caught-cheating-on-naplan-20130216-2ek6p.html

Torgesen, J. K., Wagner, R., Rashotte, C., Alexander, A., & Conway, T. (1997). Preventative and remedial interventions for children with severe reading disabilities. Learning Disabilities: A Multidisciplinary Journal, 8, 51-61.

Torgesen, J.K. (1998, Spring/Summer). Catch them before they fall: Identification and assessment to prevent reading failure in young children. American Educator. Retrieved from http://www.ldonline.org/article/225/

Vellutino, F. R., Scanlon, D. M., Sipay, E. R., Small, S. G., Pratt, A., Chen, R., & Denckla, M. B. (1996). Cognitive profiles of difficult to remediate and readily remediated poor readers: Early intervention as a vehicle for distinguishing between cognitive and experiential deficits as basic causes of specific reading disability. Journal of Educational Psychology, 88, 601-638.

Wade, B., & Moore, M. (1993). Experiencing special education. Buckingham: Open University Press.

Wanzek, J., Vaughn, S., Scammacca, N.K., Metz, K., Murray, C.S., Roberts, G., & Danielson, L. (2013). Extensive reading interventions for students with reading difficulties after Grade 3 Review of Educational Research 8(2), 163-195.

Watson, A. (1998). Potential sources of inequity in teachers' informal judgements about pupils' mathematics. Paper presented at Mathematics Education and Society: An International Conference, Nottingham, September. University of Oxford Department of Educational Studies. Retrieved from http://www.nottingham.ac.uk/csme/meas/papers/watson.html

Assessment beyond NAPLAN Dr Kerry Hempenstall

Broad scale assessment at national level (NAPLAN) is valuable in helping answer the question of how well we are doing across the nation in literacy. There are limitations to the national testing regime when the only tasks included are those intended to assess reading comprehension. It is unquestionably a major function of reading, but not the only important component.

The assessment of other critical components can supply valuable information not available in the NAPLAN process. For example, other forms of assessment can assist in the identification and management of students at-risk even before reading instruction commences. They can also help identify those making slow progress at any year level. This is especially important given the usually stable learning trajectory from the very early stages. If specific interventions are implemented, appropriate reading assessment can provide on-going information about the effectiveness of the chosen approach. There is an important question implicit in this potentially valuable activity. What sorts of assessment are likely to be most beneficial in precluding reading pitfalls and enhancing reading success? In this submission, the emphasis is directed towards assessment of those aspects of reading that have been identified by research as critical to reading development. These other forms of data collection may be made by teachers and other education-oriented professionals such as educational psychologists and speech pathologists.

Assessing literacy in Australia

The attainment of high levels of literacy in Australia remains a distant objective, apparently no closer now than in the past, despite the investment of huge sums on smaller class sizes and various instructional initiatives (Leigh & Ryan, 2008). Until recently, national assessment results have not been available in Australia, as they are in the USA through their National Assessment of Educational Progress (NAEP; Koretz, 1992), a program that has measured the reading of students in years 4, 8, and 12 since 1992. An absence of explicit, regularly collected national data has made it difficult to be precise about the extent of literacy development across the nation.

The Australian Government House of Representatives Enquiry (1993) estimated that between 10-20 % of students finish primary school with literacy problems. More recently it was reported that the remedial program known as Reading Recovery is provided to on average 40 to 50 % of Year 1 students (Office of the Victorian Auditor General, 2003). Concern has been expressed that after their Year 3 at school students with reading problems have little prospect of adequate progress (Australian Government House of Representatives Enquiry, 1993). Providing additional foundation for that fear was a Victorian study (Hill, 1995) that noted little discernible progress in literacy for the lowest 10% between Year Four and Year Ten. Nationally, according to the Australian Council for Educational Research, more than 30 % of Australian children entering high school (mainly in government and Catholic schools) cannot read or write properly (Hill, 2000). This figure of 30% is also reported by Louden et al. (2000), and Livingston (2006). Almost half of all Australians aged 15-74 years have literacy skills below the minimum level needed to manage the literacy demands of our modern society (Australian Bureau of Statistics, 2008).

In contrast to these alarming figures, government pronouncements on literacy success are usually more positive. In the recent NAPLAN national assessment of students in Year 3, 5, 7, and 9, approximately 90% of students reportedly achieved the required minimum standards (MCEETYA, 2008). Unfortunately, the benchmarks were not made transparent, and hence it is difficult to reconcile these findings with other assessments described above. Knowing what constitutes minimum standards is vital, given the marked variability displayed in the previous national and state assessment schemes that the NAPLAN replaced.

A weakness of such opaque data is the potential for benchmarks to be manipulated to show governments of the day in the best possible light. There are examples in which benchmarks have been so low as to be at the level of chance. For example, when four multiple choice items constitute the response set for students, a 25% mark could be obtained by chance alone. Surely benchmarks would never be so low that chance alone could produce a proficiency level?

"In 2006, the results needed to meet national benchmarks for students in Years 3, 5 and 7 ranged from 22% to 44%, with an average of less than 34%. Year 3 students needed to achieve only 22% for reading, 39% for numeracy, and 30% for writing to be classified as meeting the minimum acceptable standard (Strutt, 2007, p.1).”

Recently in Great Britain (Paton, 2008), the Assessment and Qualifications Alliance exam board admitted that standards had been lowered to elevate scores in 2008. In one exam paper, C grades (a good pass) were awarded to pupils who obtained a score of only 20%. Over recent years in the USA, eight states had their reading and/or maths tests become significantly easier in at least two grades (Cronin, Dahlin, Adkins, & Gage Kingsbury, 2007). The report, entitled The Proficiency Illusion, also found that recent improvements in proficiency rates on US state tests could be explained largely by declines in the difficulty of those tests.

Parental concerns about literacy are becoming increasingly evident. In the Parents’ Attitudes to Schooling report (Department of Education, Science and Training, 2007), only 37.5% of the surveyed parents believed that students were leaving school with adequate skills in literacy. There has been an increase in dissatisfaction since the previous Parents' Attitudes to Schooling survey in 2003, when 61% of parents considered primary school education as good or very good, and 51% reported secondary education as good or very good. Recent reports in the press suggest that employers too have concerns about literacy development among young people generally, not simply for those usually considered to comprise an at-risk group (Collier, 2008).

If community interest in literacy has been sparked, and there is some concern about the validity of the national broad scale assessment model, it is important for educators to offer guidance about high quality assessment. Part of the current literacy problem can be attributed to educators because they have not offered this high quality assessment in their schools to monitor progress. There has been a tendency to rely on informal assessment, such as through the use of unhelpful techniques like miscue analysis (Hempenstall, 1998), and the perusal of student folios (Fehring, 2001). If every teacher did implement a standard, agreed upon assessment schedule, based upon the current evidence on reading development, then there would be no real need for national assessment. Data would be comparable across the nation, based upon an similar metric.

It is recognised that literacy assessment itself has little intrinsic value; rather, it is only the consequences flowing from the assessment process that have the potential to enhance the prospects of those students currently struggling to master reading. Assessment also allows for the monitoring of progress during an intervention, and evaluation of success at the end of the intervention; however, the initial value relates to the question of whether there is a problem, and if so, what should be done. What should be done is inevitably tied to the conception of the reading process, and what can impede its progress. How do educationists tend to view the genesis of reading problems?

Perceptions of literacy problems and causes

Alessi (1988) contacted 50 school psychologists who, between them, produced about 5000 assessment reports in a year. The school psychologists agreed that a lack of academic or behavioural progress could be attributed to one or more of the five factors below. Alessi then examined the reports to see what factors had been assigned as the causes of their students’ educational problems.

1. Curriculum factors? No reports.

2. Inappropriate teaching practices? No reports.

3. School administrative factors? No reports.

4. Parent and home factors? 10-20% of reports.

5. Factors associated with the child? 100%.

In another study this time surveying classroom teachers, Wade and Moore (1993) noted that when students failed to learn 65% of teachers considered that student characteristics were responsible while a further 32% emphasised home factors. Only the remaining 3% believed that the education system was the most important factor in student achievement, a finding utterly at odds with the research into teacher effects (Cuttance, 1998; Hattie, Clinton, Thompson, & Schmidt-Davies, 1995).

This highlights one of the ways in which assessment can be unnecessarily limiting in its breadth, if the causes of students’ difficulties are presumed to reside solely within the students, rather than within the instructional system. Assessment of students is not a productive use of time unless it is carefully integrated into a plan involving instructional action.

When the incidence of failure is unacceptably high, as in Australia, then an appropriate direction for resource allocation is towards the assessment of instruction. It can only be flawed instruction that intensifies the reading problem from a realistic incidence of reading disability of around 5% (Brown & Felton, 1990; Felton, 1993; Marshall & Hynd, 1993; Torgesen, Wagner, Rashotte, Alexander, & Conway, 1997; Vellutino et al., 1996) to that which we find in Australia of 20 - 30% (see earlier). A tendency can arise for victim blame. "Learning disabilities have become a sociological sponge to wipe up the spills of general education. … It's where children who weren't taught well go (p.A1)" (Lyon, 1999).

Though it is not the focus of this submission, there is an increasing recognition that an education system must constantly assess the quality of instruction provided in its schools, and that it should take account of the findings of research in establishing its benchmarks and policies. “Thus the central problem for a scientific approach to the matter is not to find out what is wrong with the children, but what can be done to improve the educational system” (Labov, 2003, p.128). The interest in the national English curriculum is an example of this emerging system interest. Up to this time, education systems in Australia have been relatively impervious to such findings (Hempenstall, 1996, 2006), lagging behind significant, if tenuous, changes in the USA with Reading First (Al Otaiba et al., 2008) and in Great Britain, the Primary National Strategy (2006).

Even allowing that the major problem for the education system lies in the realm of instruction, particularly in the initial teaching of reading, individual student assessment remains of value. It is, of course, necessary as a means of evaluating instructional adequacy. Beyond that, there is great potential value in the early identification of potential reading problems, in determining the appropriate focus for instruction, in the monitoring of progress in relevant skill areas, and with the evaluation of reading interventions. It is the assumption in this paper that decisions about assessment should be driven by up-to-date conceptions of the important elements in reading development.

Issues in reading development that could guide assessment

In the largest, most comprehensive evidenced-based review ever conducted of research on how children learn to read the National Reading Panel (NRP; National Institute of Child Health and Human Development, 2000) presented its findings. For its review, the Panel selected methodologically sound research from the approximately 100,000 reading studies that have been published since 1966, and from another 15,000 earlier studies.

The specific areas the NRP noted as crucial for reading instruction were phonemic awareness, phonics, fluency, vocabulary, and comprehension. Students should be explicitly and systematically taught:

  1. Phonemic awareness: The ability to hear and identify individual sounds in spoken words.
  2. Phonics: The relationship between the letters of written language and the sounds of spoken language.
  3. Fluency: The capacity to read text accurately and quickly.
  4. Vocabulary: All the words students must know to communicate effectively.
  5. Comprehension: The ability to understand what has been read.

For children in pre-school and in their first year of formal schooling, the Panel found that early training in phonemic awareness skills, especially blending and segmenting, provided strong subsequent benefits to reading progress. It further recommended that conjoint phonemic awareness and phonics emphases should be taught directly, rather than incidentally, as effective instruction in both skills leads to strong early progress in reading and spelling.

The Panel’s emphasis on these five elements is also consonant with the findings of other several major reports, such as those of the National Research Council (Snow, Burns, & Griffin, 1998), the National Institute for Child Health and Human Development (Grossen, 1997), the British National Literacy Strategy (Department for Education and Employment, 1998), and recently in the Rose Report (Rose, 2006) and the Primary National Strategy (2006).

In 2006, the Primary Framework for Literacy and Mathematics (Primary National Strategy, 2006) was released, updating its 1998 predecessor, and mandating practice even more firmly onto an evidence base. In particular, it withdrew its imprimatur from the 3-cueing system (Hempenstall, 2003), and embraced the Simple View (Hoover & Gough, 1990) of reading that highlights the importance of decoding as the pre-eminent strategy for saying what’s on the page, and comprehension for understanding that which has been decoded. Under the 3-cueing system, making meaning by any method (for example, pictures, syntactic, and semantic cues) was considered worthwhile, and, for many protagonists, took precedence over decoding as the prime strategy (Weaver, 1988).

The new 2006 Strategy mandates a synthetic phonics approach, in which letter–sound correspondences are taught in a clearly defined sequence, and the skills of blending and segmenting phonemes are assigned high priority. This approach contrasts with the less effective analytic phonics, in which the phonemes associated with particular graphemes are not pronounced in isolation (i.e., outside of whole words). In the analytic phonics approach, students are asked to analyse the common phoneme in a set of words in which each word contains the phoneme being introduced (Hempenstall, 2001). The lesser overall effectiveness of analytic phonics instruction may be due to a lack of sufficient systematic practice and feedback usually required by the less able reading student (Adams, 1990).

In Australia, the National Enquiry into the Teaching of Literacy (Department of Education, Science, and Training, 2005) recommendations exhorted the education field to turn towards science for its inspiration. For example, the committee argued strongly for empirical evidence to be used to improve the manner in which reading is taught in Australia.

In sum, the incontrovertible finding from the extensive body of local and international evidence-based literacy research is that for children during the early years of schooling (and subsequently if needed), to be able to link their knowledge of spoken language to their knowledge of written language, they must first master the alphabetic code – the system of grapheme-phoneme correspondences that link written words to their pronunciations. Because these are both foundational and essential skills for the development of competence in reading, writing and spelling, they must be taught explicitly, systematically, early and well (p.37).

Research supporting an early emphasis on the code for both assessment and instruction?

Even though it is comprehension that is the hallmark of skilled reading, it is not comprehension per se that presents the major hurdle for most struggling young readers. There is increasing acknowledgement that the majority of reading problems observed in such students occur primarily at the level of single word decoding (Rack, Snowling, & Olson, 1992; Stanovich, 1988a; Stuart, 1995; Vellutino & Scanlon, 1987), and that in most cases this difficulty reflects an underlying struggle with some aspect of phonological processing (Bradley & Bryant, 1983; Bruck, 1992; Lyon, 1995; Perfetti, 1992; Oakhill & Garnham, 1988; Rack et al., 1992; Share, 1995; Stanovich, 1988a, 1992; Vellutino & Scanlon, 1987; Wagner & Torgesen, 1987). In the Shaywitz (2003) study, 88 percent of the children with reading problems had phonologically-based difficulties. Lovett, Steinbach, and Frijters (2000) summarise neatly this emphasis. “Work over the past 2 decades has yielded overwhelming evidence that a core linguistic deficit implicated in reading acquisition problems involves an area of metalinguistic competence called phonological awareness” (p.334).

Unless resolved, phonological problems predictably impede reading development, and they continue to be evident throughout the school years and beyond (Al Otaiba et al., 2008). A study by Shankweiler, Lundquist, Dreyer, and Dickinson (1996) provided some evidence for the fundamental problem area. Their study of Year 9 and Year 10 learning disabled and low to middle range students found significant deficiencies in decoding across the groups, even among the average students. They argued for a code-based intervention as an important focus. They also noted that differences in comprehension were largely reflecting levels of decoding skill, even among senior students, a point echoed by Simos et al. (2007) in their magnetoencephalographic study, and Scammacca et al. (2008) in their meta-analysis. Shankweiler and colleagues (1999) also found that decoding, assessed by reading aloud a list of non-words (e.g., skirm, bant), correlated very highly with reading comprehension -- accounting for 62% of the variance.

A number of similar studies involving adults with reading difficulties have revealed marked deficits in decoding (Bear, Truax, & Barone, 1989; Bruck, 1990, 1992, 1993; Byrne & Letz, 1983; Perin, 1983; Pratt & Brady, 1988; Read & Ruyter, 1985; cited in Greenberg, Ehri, & Perin, 1997). In the Greenberg et al. (1997) study with such adults, performance on phonologically-based tests resembled those of children below Year Three. Even the very bright well-compensated adult readers acknowledged that they had laboriously to remember word shapes (an ineffective strategy), had little or no idea how to spell, and were constantly struggling to decode new words, especially technical terms related to their occupations.

The emphasis on decoding is not to say that difficulties at the level of comprehension do not occur, but rather, that for many students they occur as a consequence of a failure to develop early fluent, context-free decoding ability. The capacity to actively transact with the text develops with reading experience, that is, it is partly developed by the very act of reading. Students who engage in little reading usually struggle to develop the knowledge of the world and the vocabulary necessary as a foundation for comprehension (Nagy & Anderson, 1984; Stanovich, 1986, 1993). “ … the phonological processing problem reduces opportunities to learn from exposure to printed words and, hence, has a powerful effect on the acquisition of knowledge about printed words, including word-specific spellings and orthographic regularities” (Manis, Doi, & Bhadha, 2000, p.325).

Schools often espouse the goal of teaching all students to read. So, they need to know how students are progressing along the way to meet this goal (Kame'enui, Simmons, & Coyne, 2000). This implies the existence of long-term reading goals, and some sort of performance benchmarks for their students. Criterion-based benchmarks supply one form of progress monitoring. They should be aligned with the skills emphasised by the National Reading Panel, and assessed regularly during the primary years at least, to provide schools with information about student progress and the appropriateness (focus, intensity, duration) of the current instruction (Coyne, Kame'enui, & Simmons, 2004).

Effective initial instruction can reduce the need for much individual assessment, but it is predicated on the provision of regular whole class monitoring. For example, at the beginning of the year all students could be provided with a reading assessment to assess overall literacy competence. A mid-year progress assessment can be used to evaluate instructional adequacy, and inform any revised instructional decisions. For those detected as being at risk, more fine-grained assessment information allows for efficient, pinpoint instruction. This group of students require elevated intensity of programming with specific short-term measurable objectives. Those short-term learning goals could be monitored at least monthly. In this system, those who struggle are observed more closely and more frequently.

 

Given the confluence of the findings of empirical research on reading instruction, it is appropriate for reading assessment to reflect this current understanding. If the five NRP elements are critical to development, then designing assessment around these five offers the best chance of detecting where something goes wrong, rather than solely that something is wrong. Of the five important elements, phonemic awareness, phonics, and fluency are lower order skills related to efficiently getting the words off the page; whereas, vocabulary and comprehension are higher order language skills associated with appreciating the meaning of the words obtained through efficient use of these well developed lower order processes.

Reading assessment that reflects the current understanding

At beginning stage:

Early reading delay is sometimes viewed as indicative of a slow starter who will catch up later; however, this is a dangerous assumption. It is based upon a belief that learning to read is as natural as learning to speak, and that immersion in interesting literature is sufficient to promote the process of development. An associated assumption that delays early identification and intervention is that children have a natural and immutable developmental trajectory that cannot profitably be hurried (Hempenstall, 2005).

Juel (1988) reported a probability of .88 that a student classified as a poor reader at the beginning of Year 1 would remain so when re-tested at Year 4, a finding echoing earlier work by Satz, Fletcher, Clark, and Morris (1981) who found that 93.9% of severely poor readers in Year 2 continued to be poor readers in Year 5. There is now a strong consensus (Al Otaiba et al., 2008; Shaywitz et al., 1999) about this nexus. The Matthew Effect (Stanovich, 1986) describes how relatively minor early deficits often broaden into a cascade of problems that intensify over a student’s career. Early identification and intervention should be paramount issues for the sake of those children who are at present needlessly exposed to crippling, extended failure.

Methods of identification

It is possible to obtain information from a wide array of domains possibly relevant to reading success, for example, perceptuo-motor development, and skills related to vision, balance, speech, handedness, self-help, language, and socialisation. One important issue is to what degree the potential components add to the predictive power of any intended assessment battery.

A second is how accurately the test(s) predict membership of the group of students who will struggle. A test or test battery, when employed with a group of children, will have a false positive rate and a false negative rate. In the former, there are children identified as at-risk who, without intervention, do not subsequently present with literacy difficulties. In the latter, there are children who do not appear as at-risk in the assessment, but who do later develop literacy problems. Scarborough (2003) refers to these occurrences as miss errors. Depending on the test(s) chosen and the cut-offs selected as the risk levels, the test(s) may be overly inclusive, identifying an unreasonably large cohort of students as at-risk. Alternatively, the results may underestimate the number of children at risk. The choice of tests needs to take these issues into account, and a higher than desirable false positive rate is considered more acceptable than a high false negative rate. Although one might waste resources on some students who didn’t really require additional assistance, at least you don’t miss many of those who do need your help.

Tests are necessary because teacher predictions are highly variable. As a group, teachers tend to predict reasonably well those students who will not experience reading difficulty, but predict much less well those students who will subsequently endure literacy problems (Rathvon, 2004).

The third requirement is that any screening test be relatively brief, and readily administered by teachers (Snow, Burns, & Griffin, 1998). The more (relevant) tests you add, the more accurate will be your predictions; however, the scarce resources available for testing all children demand brief assessment times. Scarborough (1998) in her review of screening tools suggested the extra work involved in administering and interpreting large test batteries is not reflected in a commensurate improvement in accuracy of screening.

The choice of tests also involves a consideration of the theoretical relevance of the tests, their soundness as instruments (validity and reliability), and their time demands. The length of some tests and batteries makes their use on large cohorts unworkable. There is usually some trade-off when selecting tests between the higher reliability of longer tests and multiple tests of the same skill, and the feasibility of test use with cohorts of students. However, a great deal of effort is being expended in constructing tests (both standardised and curriculum-based) that are brief, and yet able to provide trustworthy and valuable information. The tests described in this paper are generally considered to be of sound construction, and have adequate validity and reliability. However, a problem with many tests involves floor effects (Rathvon, 2004). When there are insufficient low-complexity items in a test, a very low score may not be adequately interpreted as an at risk score when standardised. Thus, a test with a high floor may fail to detect serious deficits. This problem occurs most frequently with the youngest children in a test’s age range. For example, a raw score of 1 out of 20 items in the Elision subtest of the Comprehensive Test of Phonological Processing (CTOPP) (Wagner, Torgesen, & Rashotte, 1999) produces a standard score of 8 for a five year old, which is less than one standard deviation from the mean (M=10, SD=3). Thus, test users need to be alert to test issues generally, and especially to the risk of even highly regarded tests having weaknesses at their lower age ranges.

The advantage of broadscale screening is that it avoids the backlog, delay, and variable accuracy of teacher judgement when individual referral is the only detection system. Individual referral is a much less reliable system, and often results in reading difficulties lying undiscovered until they are well advanced, resistant to intervention, and have broadened to include many curriculum areas (Coyne, Kame'enui, & Simmons, 2004).

Schedule for assessment

A further question involves when best to schedule these screens. The later the screen, the more accurate the predictions tend to be (Scarborough, 1998). However, screening in mid Year 1, while it may provide a clearer picture of who is in need of intervention, allows more than a year to be wasted and failure to become entrenched. On the other hand, screening during the pre-school year may over-identify those who have simply had little literacy experience. One compromise is to schedule screening on several occasions during the primary years, most frequently in the early years (Rathvon, 2004). Work is still progressing in endeavouring to make accurate predictions at the preschool and beginning reading stage.

Again there are several broad formats. One might initially screen all incoming beginning students, for example. Those who fall below a predetermined criterion are allocated additional or more intensive assistance, and subsequently re-screened. The criterion or cut-off is often the lowest quartile (25%), but some suggest selecting the lowest 10%. The choice depends to some degree on the resources available to address the resultant demands on the school resources. If one selects the lowest 25% there will be a larger cohort for whom intervention will need to be supplied. The outcome will entail a larger false positive rate, but a lower false negative rate. A larger group will include students who didn’t really need extra help, as it eventuates, but it shouldn’t miss many of those who really do. If the resources are available, this more inclusive criterion is a good option. If the lowest 10% are chosen, fewer false positives but more false negatives will arise. So, this system may ignore some students who subsequently display reading difficulties.

A variant is to employ a two-tier screen in which those detected in the process above are provided with more detailed assessment, involving a more comprehensive set of measures designed to specify with more precision the optimal area(s) for prophylactic intervention. A child’s low score on a screening test does not describe the nature of the problem, but rather serves as a beacon that there is an issue of concern.

A third model treats all students as the target group. Everyone is screened with, for example, a fluency measure that establishes their baseline attainment in February. All the students participate in an evidence-based literacy program, and their growth is monitored with similar, but time-sensitive, scheduled progress (formative) assessments, usually three times over the year. Those students not displaying acceptable growth curves are assigned additional support.

California’s Reading First Plan (California Department of Education, 1999) is well regarded as a model of careful, regular, evidence-based assessment. The sequence they adopted is summarised in Table 1.

 

Table 1 California’s Reading First Plan

Phoneme Awareness: Mid-year for Prep; End of Grade 1 (if needed); only if needed for Grades 2 and 3

Tasks: Deletion: Initial and Final Sounds, Phoneme Segmentation, Counting Syllables.

Beginning Phonics: Late Prep; only if needed for Grades 1, 2, and 3;

Tasks: Alphabet Names, Consonant Sounds, Short Vowel Sounds.

Phonics: Every 4 to 6 Weeks for Grade 2; only if needed for Grade 3;

Tasks: Word Study, Decoding, Early spellings.

Oral Reading Fluency: Early Grade 1, then 3 to 6 times per year for Grades 2 and 3;

Tasks: Timed Fluency – Words Correct Per Minute (WCPM).

Reading Comprehension: Every 8-10 Weeks, Grades 1, Every 6-8 Weeks, Grades 2 and 3;

Tasks: Main idea, Author’s point of view, Analysis, and Inference.

Vocabulary: Every 8-10 Weeks, Grades 1; Every 6-8 Weeks, Grades 2 and 3;

Tasks: Antonyms, Synonyms, Multiple Meanings, Context Meanings.

This approach is sometimes called dynamic or interactive assessment, and sometimes the test-teach-test model. It is based upon the idea that more may be learned about children's cognitive development through the assessment of what they can achieve after teaching, rather than solely assessing unassisted performance, as in static tests. It is considered by Tzuriel (2000) to more accurately reflect the learning potential of children than do snapshot tests. It also forms the basis for the approach to special education known as Response to Intervention (RTI; Gresham, 2001). Children displaying signs of failing in the early grades receive scientifically based instruction as soon as possible. Special education services focus on those who, even with these services, are not successful. These have been described as treatment resisters (Torgesen, 2000). The focus of RTI is on responding to the instructional challenges caused by the disability, rather than solely giving tests to document the failure of the student.

According to Carnine (2003), the Response to Intervention model includes five major steps:

First, criteria in the beginning grade determine if any child exhibits a significant difference between actual and expected rate of learning in one or more of the academic domains included in the definition of specific learning disabilities

Second, develop a plan to provide an evidence-based intervention. Ensure that the teacher receives training sufficient to fully implement the intervention.

Third, monitor and document the child’s progress, and regularly report this information to parents.

Fourth, if the child is not progressing at a desired rate, determine if the intervention is being implemented with fidelity -- and if not, provide additional assistance to the teacher.

Fifth, lack of progress over an agreed limited period of time leads to a full child-centred evaluation conducted by a team. In the USA, this process could lead to identification of the child as having a specific learning disability, and subsequently to the provision of special education services.

A crucial and far reaching element in establishing a systematic assessment plan in a school is the likely casualty rate. The figures cited earlier suggest 20-30% of students currently struggle to varying degrees. It is also considered by many that the most common approaches to the initial teaching of reading are suboptimal, and are themselves responsible for a high proportion of the failure rate (National Inquiry into the Teaching of Literacy, 2005). Given the number of students who struggle to master reading, efficiency in the provision of initial teaching and subsequent support becomes very important for education systems (Rayner, Foorman, Perfetti, Pesetsky, & Seidenberg, 2001). There are several components of effective whole-system or whole-school approaches. The first is attention to evidence as a means of determining the most effective approaches. Then, adequate time must be assigned to the task of providing initial reading instruction, although not all students require the same level of direct teacher input. With improvements in intensity and program quality, a reduction in the number of students requiring significant 1:1 teacher time then allows additional time to be provided for the seriously struggling students. This circumstance can eventuate when initial instruction reflects effective, research-supported approaches, thereby producing fewer casualties, and enabling the school costs of providing intensive support to be maintained at realistic levels. The third component entails a recognition that, for some students, the duration of intervention may be markedly extended beyond the average for literacy to develop successfully (Torgesen, 2000).

An early screen protocol

Over time it has become apparent that the strongest predictors of success in beginning reading are a knowledge of letter-sound correspondences (Chall, 1967; Stage, Sheppard, Davidson, & Browning, 2001) and phonemic awareness (Torgesen, 1998). This provides a theoretical rationale for focussing assessment on these areas initially.

Torgesen (1998) suggests a screening procedure involving: 1) a test of knowledge of letter names or sounds, because letter knowledge continues to be the best single predictor of reading difficulties; and 2) a test of phonemic awareness. Torgesen’s research indicates that, individually, knowledge of letter names is the stronger predictor for children in their first year, and knowledge of letter-sounds is stronger for first graders. McBride-Chang (1999) considers letter-sound knowledge to be more closely related to reading skills than is a grasp of letter names, because of the stronger phonological basis for letter-sound knowledge. Thus, assessing letter names has predictive value only because it is a marker for a range of previous useful literacy experiences rather than a direct cause of progress. However, letter-sound knowledge appears to have a causal rather than merely correlational relationship to reading progress (Levin, Shatil-Carmon, & Asif-Rave, 2006).

As phonemic awareness is thought to involve a developmental sequence, the decision as to which form of test to employ for a student cohort assumes importance. For example, it is recognised that blending, segmenting, and deletion are quite difficult tasks for children before and during their first year of school (Schatschneider, Francis, Foorman, Fletcher, & Mehta, 1999). Tests in which few students can achieve success or tests in which most students are near ceiling are of little use as screening devices.

In a longitudinal study of 499 children from kindergarten through Grade 3 (Vervaeke, McNamara, & Scissons, 2007), an accuracy figure of 80% was obtained when kindergarten assessment of phonological awareness and letter-sound correspondence was compared to their Grade 3 reading achievement. The false negative and false positive rates were each 12%, representing encouraging predictive capacity over a significant period of time.

Letter knowledge

One letter knowledge test is the Letter Identification subtest of the Woodcock Reading Mastery Test-Revised NU (Woodcock, 1998). It presents letters in several different fonts for which either the sound or the name is scored as correct. Its use of different fonts appears to be intended to enable the assessment of the concept of sound-symbol relationship, not simply the association between one letter-shape and its name/sound.

The Comprehensive Inventory of Basic Skills -Revised (Brigance, 2000) has several useful subtests. Visual discrimination of upper and lower case letters, Recitation of the alphabet, Reading upper and lower case letters, Printing upper and lower case letters in alphabetic sequence, and, Printing upper and lower case letters as dictated.

The Neale Analysis of Reading Ability (Revised) (Neale, 1988) has a supplementary test that assesses the names and sounds of the alphabet.

Good and his colleagues (Good & Kaminski, 2002) have established performance-based benchmarks using the freely available Dynamic Indicators of Basic Early Literacy Skills (DIBELS). The tests relevant to this screening task are Letter Naming Fluency and Initial Sound Fluency. Note that these tests are timed, so they add a component of speed along with power – efficiency along with knowledge. Employing fluency in the measurement of subword skills (e.g., letter names/sounds) has become of increasing interest (Speece, Mills, Ritchey, & Hillman, 2003) because of the significance of automaticity as a quality beyond mastery.

The DIBELS measures are also very brief, and easy to administer. Letter Naming Fluency involves a sheet with upper and lower-case letters, and students name as many letters as possible in 1 min. Fewer than 2 letters in 1 min at preschool or early first year at school is considered at-risk, between 2 and 7 constitute some risk, and 8 or more is classed as low risk.

Phonemic awareness

In the DIBELS Initial Sound Fluency, students are shown (for 1 minute) a series of pages containing 4 pictures. (Pointing to the pictures) This is: tomato, cub, plate, doughnut. Which picture begins with /d/? Fewer than 4 initial sounds correct in 1 minute at preschool or early first year at school is considered at-risk, between 4 and 7 constitute some risk, and 8 or more is classed as low risk.

A similar free system called AIMSweb is available for download (The Psychological Corporation, 2007). It includes subtests for Phoneme Segmentation, Letter Naming Fluency, and Letter Sound Fluency.

Thus, one direction is early screening of all students using a couple of simple tests. When more detailed testing is required, what areas are most fruitfully explored? It seems appropriate to focus on the areas deemed by the National Reading Panel (2000) to be pivotal to reading development.

More detailed testing of each of the NRP five instructional emphases

NRP instructional recommendation 1: Phonological (or phonemic) awareness

There is strong consensus among researchers that phonemic awareness is a robust predictor of future reading progress, markedly better than is intelligence (Stanovich, 1991). As this awareness is also thought by many (Adams, 1990), but not all (Castles & Coltheart, 2004), to be a major causal factor in early reading progress assessment of current levels allows both a prediction of a child’s likely progress in the absence of appropriate intervention, and a direction for any intervention to take.

Phonological (or phonemic) awareness is an auditory skill enabling the recognition that the spoken word consists of individual sounds. It appears to follow a developmental sequence: from simple (Do cat and comb begin with the same sound?) to complex (blending, and then segmenting). A study by Schatschneider et al. (1999) suggests that phonemic awareness is a unitary construct, but that its development is best charted as a sequence of tasks from simple phonological tasks, like rhyming to more complex appreciation and manipulation of phonemes, as in elision, blending, and segmenting.

In a huge study (Hoien, Lundberg, Stanovich, & Bjaarlid, 1995), initial-phoneme and final-phoneme matching tasks (such as assessed by the TOPA: Test of Phonological Awareness (Torgesen & Bryant, 1994)) were by far the most potent phonological predictors of early reading acquisition. There are a number of screening tests available, but fewer with norms, the TOPA being one that has an age range of 5.0 - 8.11 years. Another advantage of this instrument is its facility for group-testing.

Another test is the Phonological Awareness Screening Test (Henty, 1993) developed in Tasmania for which the author has been attempting to obtain normative data. The Sutherland Phonological Awareness Test–Revised (SPAT-R; Neilson, 2003) has norms (Australian) for Years P-3. The Lindamood Auditory Conceptualization Test (Lindamood & Lindamood, 1979) has norms for Years P-12. The Rosner Test of Auditory Analysis Skills (Rosner, 1975) is a 13 item test with norms for Years P-3. The Yopp-Singer Test of Phoneme Segmentation (Yopp, 1995) is a brief test for P-1 students. Informal un-normed tests are available in A Sound Way (Love & Reilly, 1995), Sound Linkage (Hatcher, 1994), Phonemic Awareness Checklist (Lewkowicz, 1980), Phonemic Awareness in Young Children (Adams, Foorman, Lundberg, & Beeler, 1998), among others.

Other phonological processes

There are at least two phonological skills besides phonemic awareness, and they are beginning to assume importance in the research literature because of their capacity to add discrimination power to test batteries (Badian, 1994; Cornwall, 1992; Felton, 1992; Hurford et al., 1993; Hurford, Schauf, Bunce, Blaich, & Moore, 1994; Spector, 1992), and because they may play a role in the development of reading beyond that contributed by phonemic awareness (Savage & Frederickson, 2005). They will often form part of the more comprehensive assessment in the two-tier screen approach.

1. Phonological recoding in lexical access.

Humans store the internal representations of words in sound form known as phonological segments. These representations need to be clearly distinguishable from other stored sound segments, or else the wrong word may be selected when, for example, one is asked to name an object presented in a picture, or a written number, or letter.

Not only must the representations be distinct, but they must be quickly and accurately accessible. Students with reading difficulties often display significant difficulty with rapidly retrieving and accessing names for visual material, even though the relevant names are known to them. The impact on reading development is that a deficit in this area will also adversely impact upon the basic processing necessary for fluent word recognition processes, and thereby reading comprehension (Wolf, Miller, & Donnelly, 2000). Savage and Frederickson (2005) found that alphanumeric naming capacity was particularly strongly associated with reading fluency.

These speed and accuracy problems may be evident even prior to experience with print. Naming speed for pictures or objects may be slow, as too, subsequently, naming of (known) numbers and letters. A number of researchers have noted the predictive power of naming-speed tasks, using pictures, numbers, and letters. Both naming speed and sight word reading depend on rapid, automatic symbol retrieval. Bowers (1995) argues that slow naming speed is specific to reading disability, and not common to children with either broad-based reading problems, or Attention-Deficit Hyperactivity Disorder.

Wolf and Bowers (2000) discuss the possibility that naming speed is independent of phonemic awareness and represents a second core deficit among some disabled readers (Bowers & Wolf, 1993; Miller & Felton, 2001; Wolf & Bowers, 1999). This issue is important because there may be a group whose phonemic awareness is developing normally, and who would be unidentified by a phonemic awareness screen, but who may subsequently have reading difficulties.

Additionally, there may be a group of students who have deficits in both phonemic awareness and rapid naming. Their dual difficulty may well lead them to be especially resistant to the standard procedures in reading instruction. Wolf and colleagues have described this as the Double Deficit Hypothesis. Identifying this group before the failure process commences is obviously worthwhile, because it enables the marshalling of resources to provide very intense (evidence-based) instruction to this targeted group.

A study by Lovett, Steinbach, and Frijters (2000) underlines the importance of recognising such treatment resisters. They noted that, when intensive phonologically-based instruction was implemented, even the Double Deficit students made progress commensurate with their less disabled single deficit peers. Without such carefully planned intervention, they tend to be the most severely disabled readers, and their difficulties are not relieved by maturation (Lovett, et al., 2000; Wiig, Zureich, & Chan, 2000).

Tests: RAN: Rapid Automatized Naming (Denckla & Rudel, 1974); BNT: Boston Naming Test (Kaplan, Goodglass & Weintraub, 1983); SNS: Symbol Naming Speed (Swanson, 1989); Picture Naming Test (Hempenstall, 1995). Wiig, Zureich, and Chan (2000) argue for pictures and colours as more suitable because of the exceptionally automatised nature of letter and number knowledge. However, others have found that, for children with well-established letter-sound recognition, a letter-naming test may be a better predictor (Manis, Doi, & Bhadha, 2000; Savage & Frederickson, 2006).

2. Phonological recoding in working memory.

The beginning reader is required to decode a series of graphemes, and temporarily order them to allow the cognitively expensive task of blending to occur. This skill has been found to be an important determinant of early reading success. It is usually assessed by digit span (oral & visual) and sentence memory tasks.

Tests:

Wechsler Intelligence Scale for Children: 4th Edition (WISC IV) (Wechsler, 2003): Digit Span subtest; Wechsler Pre-School and Primary Scale of Intelligence- III (WPPSI-III) (Wechsler, 2002): Sentences; Stanford-Binet: Fifth Edition (Roid, 2003): Memory subtests; Comprehensive Inventory of Basic Skills-Revised (Brigance, 2000): Sentence Memory.

Assessing all three processes

The Comprehensive Test of Phonological Processing (CTOPP) (Wagner, Torgesen & Rashotte, 1999) assesses all three phonological processes: phonological awareness, rapid naming, and phonological memory. The CTOPP is designed to identify individuals from prep to tertiary level whose reading would benefit from development of their phonological skills. One version, developed for children aged 5 and 6 has seven core subtests and one supplementary test. The second version (ages 7 to 24 years) contains six core subtests and six supplementary tests. Individual administration requires about 30 minutes for the core subtests. The CTOPP authors argue for three potential classroom uses: to provide a screening test for students who may not be developing their phonological abilities; to indicate any student’s areas of strength and weakness among those processes; and, to measure progress in phonological processes when intervention programs are in place. The subtests are Elision, Blending Words, Sound Matching, Memory for Digits, Nonword Repetition, Rapid Color Naming, Rapid Digit Naming, Rapid Letter Naming, Rapid Object Naming, Blending Nonwords, Phoneme Reversal, Segmenting Words, and Segmenting Nonwords.

Training other phonological processes?

Even though rapid naming tasks assist in the prediction of early reading success, there is as yet little evidence that directly training those tasks improves reading (Spear-Swerling, 1998). That is not to say that such efforts can never be fruitful. Wolf, Miller, & Donnelly (2000) have developed a program (Rave-O) designed to directly address the processing deficits they consider produce impediments to reading fluency. The RAVE-O program is not a stand-alone approach, but is integrated with a phonological analysis and blending strategy based upon Reading Mastery I/II Fast Cycle (Engelmann & Bruner, 1988). The additions emphasise orthographic pattern recognition, semantic development, and retrieval strategies. Independent evaluations are as yet incomplete.

Several studies have noted improvement in lexical access following phonemic awareness intervention (Beck, Perfetti, & McKeown, 1982; McGregor & Leonard, 1995). Gillam and Van Kleeck (1996) reported a study in which pre-school aged children with speech and language disorders improved both in phonemic awareness and phonological working memory following a phonemic awareness training program. Further, they noted that children with poor initial phonological working memory were as responsive to the intervention as were those with better phonological working memory. No studies thus far have supported the value of directly teaching naming or short term memory skills.

Elbro, Nielsen, and Petersen (1994) argue that poor phonological representations of words form the core deficit in disabled readers. In this view, lexical access and working memory are restricted not because of specific modular deficits in these processes, but rather because what is sought in the lexicon, or to be held in working memory, is lacking in readily distinguishing features. They noted the confusion of similar sounding words, and the less distinct word-naming in such readers. This view also finds support in a study by Eden, Stein, Wood, and Wood (1995a; 1995b). The phonological representation explanation allows for the possibility that improved phonemic awareness may lead to an assessed improvement in one or more of these other phonological processes. In fact, Rubin, Rottella, Schwartz, and Bernstein (1991) found that training Year 3 children in phonemic awareness had a significantly beneficial effect on the picture naming speed of both the good and poor readers.

Interpreting process assessment

Teachers may anticipate that students with difficulties solely in phonological awareness tasks are likely to require additional care in the teaching of decoding skills, while those with problems solely with naming speed may be expected to require assistance in whole word recognition, and careful attention to fluency development. As noted earlier, Wolf and Bowers (2000) argue that students who have difficulty with both phonological awareness tasks and naming speed tasks are very likely to be more resistant to reading instruction than are those with a problem in one area only. Schools can then prepare for intensive assistance over a longer period of time (Torgesen, Wagner, & Rashotte, 1994) with these students -- too often efforts are only irregularly scheduled and prematurely discontinued for those students in greatest need. Progress may be slow and hard earned, but attention to detail in instruction and vastly increased opportunities for practice can make a great difference to the prognosis. The lesson to be learned from assessment of student’s phonological processing is not about identifying learner characteristics to account for lack of progress, but rather to assist the discerning of which students demand our cutting-edge best interventions.

Assessment when reading instruction has commenced: Word level processes

NRP instructional recommendation 2. Phonics

Phonemic awareness becomes important when beginners are faced with the challenge of making sense of the English alphabetic system of writing. The phonological skills of blending and segmenting act upon the knowledge of letter-sound correspondences to enable the decoding of the written word. The facility with which students can do so to decode words not before seen is a necessary step on the way to effortless fluent reading. The decoding of non-words is considered the most appropriate measure of this process (Hoover & Gough, 1990; Siegel, 1993; Wood & Felton, 1994). While it may appear to be a task only obliquely related to reading, the measure ensures that memory for words and contextual cues can be ruled out as explanations when the non-words are read accurately. Non-word decoding also correlates very highly with reading comprehension (Shankweiler, Lundquist, Dreyer, & Dickinson, 1996).

Share (1995) argued that students must achieve a certain level of facility with decoding before a self-teaching mechanism allows them to make continuous independent progress from that stage, eventually employing for the most part the orthographic strategy that enables rapid, accurate, effortless reading. This self-teaching view is supported by several studies (Bowey & Muller, 2005; Nation, Angell, & Castles, 2007; Share, 1999; 2004) highlighting the centrality of decoding to reading development, most particularly at the early stages. These sought-after orthographic strategies can only be developed through multiple examples of success in decoding phonologically (Ehri, 1998; Share & Stanovich, 1995).

Some research using brain imaging techniques (Joseph, Noble, & Eden, 2001; Gaillard et al. 2006; Pugh et al., 2002; Turkeltaub et al. 2004) has added to our understanding of this link. It appears that the left brain’s parieto-temporal region is employed in decoding (sounding-out), and in good readers this area is very active during reading. In struggling readers, there is little activity in the left hemisphere but considerably more in the less helpful right hemisphere (Simos et al., 2002).

When beginning readers have decoded a word correctly a number of times, they develop a neural model that is an exact replica of the printed word, reflecting the word’s pronunciation, spelling, and meaning. This internal representation is maintained in the occipito-temporal region of the left hemisphere. Subsequent recognition of that word becomes automatic, taking less than 150 milliseconds (less than a heartbeat). The development of orthographic processing, the key to fluent reading, depends upon the occipito-temporal region. However, the occipito-temporal region does not assume responsibility for the task without first the parieto-temporal region regularly being engaged (Richards et al., 2006; Shaywitz et al., 2004).

On average, from 4-14 accurate sounding-outs (Apel & Swank, 1999) will create the firm links necessary, although some children may require many times that number (Lyon, 2001; Swanson, 2001a) to facilitate the growth of connections between those regions. Not all children have a strong phonological talent, and there may be both genetic and environmental influences to create these individual differences.

The degree to which students are then able to use their developing parieto-temporal region in the reading task can be assessed with the Word Attack subtest, Woodcock Reading Mastery Tests-Revised (WRMT; Woodcock, 1998); with the Pseudoword Decoding subtest from the Wechsler Individual Achievement Test-II (WIAT-II; Wechsler, 2001); DIBELS Nonsense Word Fluency (Good & Kaminski, 2002); Phonemic Decoding Efficiency (PDE) subtest of Test of Word Reading Efficiency (TOWRE; Torgesen, Wagner, & Rashotte, 1999); or the Castles Non-Word List (Castles & Coltheart, 1993, in press). This latter test has been modified and renormed, and is published in this issue.

In the DIBELS Nonsense Word Fluency (NWF), the student in mid prep to end of 1st Year reads aloud a collection of short nonsense words as quickly as possible for one minute. NWF below 5 is considered at risk, between 5 and 12 at some risk, and 13 or more at low risk.

 

The TOWRE Phonemic Decoding Efficiency (PDE) subtest measures the number of pronounceable printed non-words that can be accurately decoded within a brief timeframe (45 seconds). It has norms from age 6 to 25 years.

 

These tests add another quality to the other tests mentioned above -- that of fluency of decoding. Fluency provides information beyond mastery, separating those who are accurate, but slow, from those for whom decoding is effortless and automatic. In Stanovich’s (2000) view, the rapid decoding of nonwords is one of the best discriminators of good and struggling readers. Comprehension is disrupted by slow word reading (Perfetti, 1985; Perfetti & Hogaboam, 1975; Perfetti & Lesgold, 1977). Words should be effortlessly identified so that word reading takes up a minimal amount of processing capacity, leaving as much as possible for understanding the text meaning (Bowers & Wolf, 1993; Campton & Carlisle 1994; Joshi & Aaron, 2002; Metsala & Ehri, 1998).

The TOWRE also contains a Sight Word Efficiency (SWE) subtest that assesses the number of real printed words that can be accurately identified within 45 seconds. The TOWRE is helpful for a number of purposes, such as in monitoring the growth in efficiency of phonemic decoding and sight word reading skills during the primary school years. It also highlights any differences between the two skills in the same student. This has implications for any intervention that may required by a student. The two subtests can be administered to a child in less than 5 minutes, and there are two parallel forms of each subtest.

It has been suggested that assessment at the level of the single word, as in lists rather than employing only authentic literature, are in some way not real tests of real reading, because it involves fractionating the reading process (Goodman, 1986). However, two studies by Landi, Perfetti, Bolger, Dunlap, and Foorman (2006) have pointed to the potential for list-type assessment to provide a purer measure of orthographic and phonological skills, because when beginning readers read words in context, they may depend on context to attempt to circumvent their inadequacies in reading unfamiliar words.

Assessment for older students

One question may be How delayed is this child's reading development? A general reading assessment will provide some information. It will provide an idea of the length of time it may take for the child to achieve a reasonable level of reading skill (i.e., to be able to adequately comprehend grade-level textbooks as a minimum outcome) given a good program, regularly and competently taught to a motivated student. Normed reading tests may be used for this purpose, bearing in mind the various problems they have in specifying precise grade levels. In the RMIT Clinic, the most commonly used general tests are the Woodcock Reading Mastery TestsRevised NU (Woodcock, 1998), the Neale Analysis of Reading Ability-Revised (Neale, 1988), the Spadafore Diagnostic Reading Test (Spadafore, 1983), and various subtests of the Comprehensive Inventory of Basic Skills-Revised (Brigance, 2000).

These tests will usually provide an indication of the student’s ability to read accurately from word lists or connected text (reading accuracy) and the capacity to make sense of that which they read (reading comprehension). Reading accuracy tests do not adequately discriminate between those students who have memorised whole words and those students who additionally have the capacity to decode words not recognised. The Woodcock has a significant advantage over the Neale because of the inclusion of a Word Attack subtest that indicates the degree to which the student can apply his phonemic awareness to the task of reading (sometimes called phonological recoding). Additionally, it is normed to an adult level. The Neale allows for testing of reading rate, an important element in a student’s progress, reflecting the level of automaticity or fluency achieved. Rate also provides information about the attentional capacity a reader has available to commit to the task of reading comprehension.

Assessing reading fluency

NRP instructional recommendation 3: Reading fluency

According to Wolf and Katzir-Cohen (2001):

In its beginnings, reading fluency is the product of the initial development of accuracy and the subsequent development of automaticity in underlying sublexical processes, lexical processes, and their integration in single-word reading and connected text. These include perceptual, phonological, orthographic, and morphological processes at the letter, letter-pattern, and word-level; as well as semantic and syntactic processes at the word-level and connected-text level. After it is fully developed, reading fluency refers to a level of accuracy and rate, where decoding is relatively effortless; where oral reading is smooth and accurate with correct prosody; and where attention can be allocated to comprehension (p. 219).

Oral reading fluency has particular relevance during the alphabetic stage of reading development because this is the phase during which self-teaching begins (Share, 1995). In the early alphabetic stage, simple letter pattern-to-sound conversion begins to provide a means of decoding unknown words, though the process is necessarily laborious (as is any new skill prior to its automatization). As they progress with their understanding of the function of the alphabet, students begin to appreciate that each time they decode an unfamiliar word its recognition subsequently becomes easier and faster. Practising decoding enables them to become adept at storing letter-patterns -- orthographic information that can dramatically hasten word recognition of these and new words (Torgesen, 1998). These are not simply visual images, but alphabetic sequences.

It is in reaching the stage of automaticity that the apparent magic of skilled reading becomes evident – whole words are recognised as quickly as are individual letters. The actual process of reading, of transforming squiggles into language, appears transparent – that is, the words seem to leap off the page and into consciousness without any noticeable effort or strategy (LaBerge & Samuels, 1974). The issue of variation in the effort required to make sense of print has been addressed by employing neuro-imaging techniques when both capable and struggling students are engaged in reading. Richards et al. (1999, 2000) noted that the poor readers used four to five times as much physical energy (oxygen, glucose) as the capable readers to complete the same phonological tasks. This difference was not observed when non-language tasks were presented. It is unsurprising that a lack of motivation to read is a serious secondary obstacle for dysfluent readers.

Oral reading fluency has been found to be strongly related to reading comprehension (Miller & Schwanenflugel, 2006; O'Connor et al., 2002; Roehrig, Petscher, Nettles, Hudson, & Torgesen, 2008). In fact, Shinn, Good, Knutson, Tilly, and Collins (1992) found that oral reading fluency in the early grades was as valid a measure of reading comprehension as of decoding ability. Others have reported correlations as high as .91 for older students (Fuchs, Fuchs, Hosp, & Jenkins, 2001). Oral reading fluency measures correlate even better with other reading comprehension tests than those same tests correlate with each other (Fuchs & Fuchs, 1992).

Both standardised and informal assessments of oral reading accuracy and rate are recommended in the National Reading Panel Report (National Institute of Child Health and Human Development, 2000). The report also recommends guided oral reading as a valuable fluency enhancing activity, yet both fluency assessment and instruction are notably absent from the reading curricula of many schools. This is unsurprising given that reading fluency is not mentioned in the English curriculum standards documents from at least three Australian states: Victoria, South Australia, and Queensland (Department of Education, Employment & Training, 2001). Perhaps the mooted National Curriculum will incorporate such an emphasis, give that the National Inquiry into the Teaching of Literacy (2005) recommended both assessment and structured teaching of reading fluency.

While suggested rates vary among writers, Howell and Nolet (2000) recommend the following benchmarks for text appropriately graded. From early Year 1 to late Year 1, the anticipated progression is from 35–50 words correct per minute; whilst from early Year 2 to late Year 2, the target is from 70–100 correct wpm; and from early Year 3 to late Year 3 the progression is from 120–140 correct wpm. A slightly different trajectory is suggested by Binder, Haughton, and Bateman (2002). They anticipate a more rapid progression throughout Year 1 reaching between 60-100 correct wpm. They also provide additional yearly expectations: Year 2–Year 3 100–120 correct wpm; Year 4–Year 5 120–150 correct wpm; Year 6–Year 8 150–180 correct wpm; and Year 9 and above 180–200 correct wpm.

The Gray Oral Reading Test–4 (Wiederholt & Bryant, 2001) is a standardized measure of oral reading allowing assessment of reading accuracy, rate, and passage comprehension for ages 6-19, as does the Neale Analysis of Reading Ability –Revised (Neale, 1988), though only for ages from 6-12 years.

An increasingly popular curriculum-based oral reading fluency measure is the DIBELS (Good & Kaminski, 2002) which provides reasonably reliable and valid indicators of skills associated with reading success from beginning to Year 6. Numerous graded passages are provided, and students read the appropriate passages orally for one minute each. The median score (words correct per minute) of the cold reading of each of three passages forms the data. Performance-based benchmarks allow the identification of children who are doing well in their reading instruction, and detects those whose response to instruction places them at risk for experiencing later reading difficulties.

They are simple, quick, and cost-effective (free) measures that are more sensitive to small changes over time than are most standardised tests. There are multiple passages for each grade level, making them easily repeatable for continuous progress monitoring. Using DIBELS, all students are expected to be assessed three times a year, while those receiving intervention are typically assessed fortnightly or monthly.

 

Another curriculum based measure is known as AIMSweb (The Psychological Corporation, 2007). It follows a similar protocol, and has multiple passages up to a Year 8 level.

 

Norms in standardised tests

An issue sometimes arises about the appropriateness of tests employing only US norms. Obviously, it would be an advantage to have local norms for all the tests we wish to use; however, the huge cost of properly norming tests is prohibitive for many local developers. There are some grounds for defending US normed tests of reading. We speak and write the same language, and, in most Australian states, we commence school at about the same age. In international comparisons (e.g., OECD, 2004; UNICEF, 2002), our average reading attainment exceeds that in the USA, perhaps because of our lower proportion of disadvantaged and non-English speaking students. The implication of this disparity is that tests using US norms may slightly flatter our students. When students do not do well on such a test, it is likely that they would actually be lower on that test using local norms than is indicated by the test manual. So, if a student, for example, scores below the 30th percentile on the TOWRE (the cut-off for being classified as at-risk), any error caused by the non-local norms is likely to lead to an underestimate of their level of difficulty.

 

Assessment of vocabulary

NRP instructional recommendation 4: Vocabulary

The significance of vocabulary in the context of reading development involves its role in underpinning reading comprehension (Beck & McKeown, 1991). Vocabulary produces correlations with reading comprehension of between .6 and .8 (Pearson, Hiebert, & Kamil, 2007), although receptive language assessment when performed at the beginning of school tends to produce lower figures, such as .38 (Scarborough, 1998).

It is acknowledged that early vocabulary development is important for later literacy, and that there are marked differences in the vocabulary levels of children at school entry. Hart and Risley (1995; 2003) observed that (on average) parents with professional jobs spoke about 2,000 words an hour to toddlers. For working-class parents, the rate averaged 1,200 words an hour, and for those receiving welfare only 600 words an hour. Hart and Risley concluded that by age 3, children receiving welfare have heard 30 million fewer words than children of professional families. A year’s preschool experience could not entirely compensate for the experiential deprivation that could occur during the first 3 years of life.

There may in the near future be the means for very early identification of language development (Swingley, 2008). Young toddlers tend to look at images or objects that are named by an adult. Through eye movement tracking while a child observes two objects (e.g., an apple and a dog) it is possible to see if the child’s eyes move to the named object. It enables a measure of a 12 month old child’s knowledge of the meaning of words that they are not yet able to articulate. This early phase is quite strongly related to success of later language tasks, and may lead to early identification and support at the optimal period for changing a child’s trajectory.

If schools are to attempt to compensate for these dramatic discrepancies noted by Hart and Risley, then vocabulary assessment needs to be included in the planning. However, there are a number of uses of the term: receptive and expressive vocabulary, oral and reading vocabulary, reading and writing vocabulary.

The National Reading Panel (National Institute of Child Health and Human Development, 2000) described numerous studies that emphasised the way in which vocabulary develops. New words taught directly in a year are typically about 300 to 500; however, the number of new words learned in a year is around 3,000 to 4,000 (Beck & McKeown, 1991). So, much of the development must be dependent upon reading. In fact, beyond Year 3 the amount of free reading is the major determinant of vocabulary growth, and the best readers may read 100 times that of the least adept (Nagy, 1998; Nagy & Anderson, 1984). So, the initial gap is inclined to widen. Why is reading such a source of vocabulary growth? Children’s books contain 50% more "rare" words than does adult television, or even the conversation of college graduates. Even popular magazines have 3 times as many opportunities for new word learning as prime-time television and adult conversation (Stanovich, 1993). Reading stories to children appears not to adequately compensate for a lack of reading experience. Listening to stories has only been shown to increase the vocabulary of above average readers (Nicholson & Whyte, 1992). Students who don’t choose to read regularly fall further behind (Matthew Effects; Stanovich, 1986). So, the extent of vocabulary knowledge is both a cause and a consequence of reading development.

It is fair to say that the field of vocabulary assessment is less well developed than some of the other dimensions of reading. A great deal of the research employed experimenter-designed tests, and hence there has not arisen a clear consensus about which type of vocabulary assessment is most helpful in relation to reading development. According to the NRP, standardized tests should only be used to provide a baseline, as they offer only a more general measure of vocabulary. For evaluating instruction, more than a single measure of vocabulary should be utilised, preferably measures associated with the teaching curriculum.

In standardized tests, one way of assessing vocabulary is to have the student select a definition for a word from a list of alternatives. Another is to ask what various words mean (WISC-IV; Wechsler, 2003). A third is to select the word that doesn’t belong in a list either spoken or written (brown, big, red, green, yellow; Brigance, 2000). In the Woodcock Reading Mastery Tests-Revised/Normative Update (1998), three subtests comprise the Word Comprehension test: Antonyms, Synonyms, and Analogies.

The most commonly employed vocabulary test is one of receptive language, using the Peabody Picture Vocabulary Test (PPVT-3; Dunn & Dunn, 1997). There is no reading involved; the task is to identify the one picture of four that matches the word spoken by the test administrator. A similar protocol is provided in the Wechsler Individual Achievement Test-II (WIAT-II; Wechsler, 2001): Receptive Vocabulary subtest. Another option is in the Vocabulary subtest of the various Wechsler scales (WISC-IV, WPPSI-III, WAIS-III: Wechsler, 2004; 2002; 1997). The Wechsler task is to provide definitions for various, progressively more complex words.

Vocabulary deficits may impede reading comprehension, but reasons for students performing poorly on a comprehension measure are not immediately obvious from only the comprehension measure. Was low attainment caused by a decoding problem, or did inattention preclude correct answers. Did the student forget the passage details because of short term memory problems, or might anxiety have interfered? Was it a metacognition failure in which the student has simply never learned strategies to aid comprehension, or was it due to a vocabulary lag? The vocabulary test can assist with this diagnosis, but is insufficient of itself.

Assessment of comprehension

NRP instructional recommendation 5: Comprehension

As the basic decoding and word recognition skills become automatised, comprehension strategies become an area of variability among students. Strategies that were adequate in simple text may become insufficient for the increasingly complex language (semantics and syntax) in the upper primary and secondary grades.

Without the automatisation of basic processes, reading comprehension progress stalls. With automatisation, students are at least able to make use of their existing oral language comprehension skills (Crowley, Shrager, & Siegler, 1997; O’Connor et al., 2002). The growth of these largely oral comprehension skills is partly dependent upon the quality and extent of oral language activities in their curriculum. The student converts print to speech (perhaps subvocally), and comprehends the speech, rather than the text directly – as in the Simple View of reading (Hoover & Gough, 1990).

However, text is not simply transcribed speech. It has its own formats, and additional comprehension strategies assume importance over the longer period of reading sophistication. Those with a history of problems will have had reduced exposure to text that hampers subsequent progress impeding their vocabulary development (Nagy, 1998), as discussed earlier.

The research into enhancing comprehension has lagged behind that for the underpinning word-level processes, though there is some agreement about a few promising components. For example, the student who interrogates the text is likely to understand more than one who passively reads it (National Reading Panel, 2000; Pressley, 2000). Useful strategies, including prediction, analyzing stories with respect to story grammar elements, question asking, image construction, and summarizing, may be intuited by some students. However, for others these strategies are highly dependent on teachers’ modelling of the process orally, and their providing multiple practice opportunities (Pressley, 2001; Swanson, 2001b)). Unfortunately, many comprehension activities in schools involve only testing students (reading a text and subsequently answering questions) rather than actually providing instruction.

Good readers are aware of why they are reading a text, gain an overview of the text before reading, make predictions about the upcoming text, read selectively based on their overview, associate ideas in text to what they already know, note whether their predictions and expectations about text content are being met, revise their prior knowledge when compelling new ideas conflicting with prior knowledge are encountered, figure out the meanings of unfamiliar vocabulary based on context clues, underline and reread and make notes and paraphrase to remember important points, interpret the text, evaluate its quality, review important points as they conclude reading, and think about how ideas encountered in the text might be used in the future. Young and less skilled readers, in contrast, exhibit a lack of such activity (e.g., Cordón & Day, 1996). (Pressley, 2000, p.548).

Given the under-developed state of research into reading comprehension, it is unsurprising that current testing instruments also have their problems. Much of the intervention research has involved experimenter-devised tests, and these have produced rather larger effect sizes than have standardised tests when evaluating the same instructional method. For the studies on question generation, the average effect size averaged about 0.90 for experimenter-written tests, which is a large effect; whereas, for standardized tests, the average effect size was small at 0.36. The pattern was similar for the multiple strategy instruction experiments in which for experimenter-written tests the average effect size was 0.88, and for standardized tests, only 0.32 (National Reading Panel, 2000). Clearly some consensus is needed about what forms of comprehension assessment are optimal for a specific given purpose.

Standardised comprehension tests are predicated on the assumption that there is a consensus on what are appropriate, progressively increasing grade levels of comprehension. However, there are many variables to cloud interpretation of results. Grade level materials can be analysed on the basis of their readability, usually utilising one or other algorithms based upon word length, word prevalence, and sentence length. However, difficulty levels of vocabulary and syntax can vary significantly across tests, and are not quantified by readability measures. Are the questions literal or inferential? Inferential questions are usually considered harder than literal questions, but both have difficulty levels along a continuum. To further complicate the issue, domain knowledge about a topic dramatically influences task success (Hirsch, 2006), as can command of English. It has also been observed that speed of comprehension is slower and test scores are lower when unfamiliar topics are read than when familiar topics arise. A weakness, then, of comprehension measures is that the methods chosen are only indirect indicators of whether the reader has got it, and to what extent. And each of the numerous and varied methods tried has had its own set of weaknesses, whether issues of validity (particularly for individual scores), external accountability, reliability, or generalisability (Pearson & Hamm, 2005). Perhaps, future brain imaging techniques will provide more insight into the process of comprehension.

Most comprehension measures occur as subtests within omnibus tests, such as the Woodcock Reading Mastery TestsRevised NU (Woodcock, 1998), the Neale Analysis of Reading Ability-Revised (Neale, 1988), Gray Oral Reading Test–4 (Wiederholt & Bryant, 2001), Spadafore Diagnostic Reading Test (Spadafore, 1983), WIAT-II (Wechsler, 2001), DIBELS Retell Fluency (Good & Kaminski, 2002), and the Comprehensive Inventory of Basic Skills-Revised (Brigance, 2000).

Of particular interest is to compare attainment on a reading comprehension task to that on a listening comprehension task. The Comprehensive Inventory of Basic Skills-Revised (Brigance, 2000) has the capacity to provide such a comparison, with its reading comprehension and listening comprehension subtests (up to Year 9). So too does the Spadafore Diagnostic Reading Test (Spadafore, 1983), and it has an advantage in that it is normed to Year 12.

The interest lies in the degree to which performance differs on the two tasks. For the average child, if reading is fluent, then the two scores should be similar. The listening comprehension task represents the language comprehension aspect in the Simple View of Reading (Hoover & Gough, 1990), and the level of reading comprehension obtained depends also upon the adequacy of the lower level processing in addition to language comprehension.

Comparing the results of listening comprehension to reading comprehension offers the capacity to define those children who have a major problem only at the level of print. They will perform well on the listening comprehension tasks, using their impressive general language skills to answer questions about a story read to them. On the reading comprehension task, however, they will do relatively poorly as their under-developed decoding skills prevent them bringing into play their well-developed general language skills.

When required to decode a passage unassisted, these students struggle, as do their garden-variety peers -- those with a non-modular, broad-based reading problem (Stanovich, 1988b). On the other hand, the garden-variety students would be expected to perform similarly on both tasks. Their reading problems are general rather than specific, and they have more than just one or two reading subskills restricting their development. Their decoding skill is commensurate with their other language skills, such that if they know the meaning of a word (or phrase, or sentence), they can comprehend it whether it is presented orally or in print. The consequence for the high listening comprehension-low reading comprehension child should be intensive assistance at the decoding level. For the low listening comprehension-low reading comprehension child, intensive assistance at both the decoding and comprehension levels is indicated.

Other possible outcomes are high listening comprehension-high reading comprehension, a result predictable from an all-round good reader; and low listening comprehension-high reading comprehension, a rare result, possibly from a student with acute attentional, hearing, or short-term memory problems. In this case, the permanence of text would allow the student to use his intact language comprehension skills; whereas, the ephemeral nature of the spoken story precludes such access. Hyperlexic students (a less common sub-group with excellent word recognition but poor reading comprehension) would not be detected by this discrepancy analysis, because their listening comprehension parallels their reading comprehension (Sparks, 1995). Hyperlexic students should not be confused with the oft seen older struggling reader who may appear to decode adequately and have only under-developed comprehension skills. These students usually have a long history of inadequate decoding skills and fluency -- a history that has compounded across domains, despite some evident improvement in decoding.

 

This listening comprehension - reading comprehension discrepancy represents an alternative definition of the group sometimes described as dyslexic; however, as with the IQ discrepancy-defined dyslexic, an issue is how great a discrepancy should be considered significant. Some have considered a two year discrepancy to be very significant (Anderson, 1991), given the extent of commonality of the tasks. However, this is clearly an arbitrary figure, its significance being higher the younger is the age of the child. This is its major value since the intervention techniques employed include systematic synthetic phonics instruction whether the difficulty is described as dyslexic or garden-variety. The dyslexic classification can, however, sensitise teachers to the possibility that dyslexic students may be more treatment-resistant (Berninger & Abbott, 1994) than garden-variety students, and may also require more intensive or extended phonologically-based instruction if their progress in a systematic synthetic phonics program appears to be unsatisfactory despite it being appropriately taught.

Assembling the results of the various assessments for a particular student leads to some interpretation of cause and possible interventions. Below is an example of the frameworks used in the RMIT Clinic to assist this judgement. It represents a way of looking at a student’s scores, in this case a student diagnosed with dyslexia, and offering intervention appropriate to the diagnosis, which in turn is based upon the assessment findings.

Insert Figure 1 about here

Interventions

One purpose for the fine-grained, domain-specific assessments described in this paper is to enable the assigning any intervention precisely to the area that is seen to be impeding student progress. This targeting enables a more efficient use of scarce school resources, and increases the likelihood of rapid progress for the student. If identified and addressed early, impediments to progress can be removed before the debilitating Matthew Effects (Stanovich, 1986) are able to force the student into the familiar and depressing downward trajectory.

Consistent with research findings (Adams, 1990; Foorman, 1995; Perfetti, 1992), good results for decoding and fluency intervention at the RMIT Clinic have come from programs with a strong synthetic phonics emphasis and involve explicit, carefully planned instructional sequences, such as Reading Mastery (Engelmann & Bruner, 1988), Teach Your Child to Read in 100 Easy Lessons (Engelmann, Haddox, & Bruner, 1983), the Corrective Reading Program – Decoding strand (Engelmann, Meyer, Carnine, & Johnson, 1999). Also enhancing reading development are programs focusing on spelling, such as Spelling Mastery (Dixon, Engelmann, Bauer, Steely, & Wells, 1998); and on writing, such as Expressive Writing (Engelmann & Silbert, 1983) and Reasoning and Writing (Engelmann & Silbert, 1991). For comprehension problems, Language for Learning (Engelmann & Osborn, 1999), and the Corrective Reading Program – Comprehension strand (Engelmann, Haddox, Hanner, & Osborne, 1989) have proved to be valuable teaching agents.

The scripted nature of these programs is a great benefit when training parents to work effectively with their children. Although this Clinic role can never hope to change the system that creates/maintains student reading failure, it does provide parents with the tools to help them compensate for the weakness of their school system.

At the system level, parents cannot be expected to be responsible for their children’s literacy development. The results of reports from Australia, USA, and Great Britain have been remarkably consistent about what is needed. The next major step involves policy reform that takes as its primary source the scientific theory of reading development and empirically validated approaches that incorporate this theory. When combined with wisely used, salient assessment instruments, the potential for the education system to enter a self-sustaining improvement cycle is very exciting.

All that is needed is for the education industry and political bureaucracies to see the light. As a first step, they might devise explicit, measurable standards, and insist upon close, transparent progress monitoring to evaluate instructional adequacy. Straightforward really!

References

Adams, M. J. (1990). Beginning to read: Thinking and learning about print. Cambridge, MA: MIT Press.

Adams, M. J., Foorman, B. R., Lundberg, I., & Beeler, T. (1998). Phonemic awareness in young children. Baltimore, MA: Brookes Publishing.

Al Otaiba, S., Connor, C., Lane, H., Kosanovich, M. L., Schatschneider, C., Dyrlund, A. K., Miller, M. S., & Wright, T. L. (2008). Reading First kindergarten classroom instruction and students' growth in phonological awareness and letter naming–decoding fluency. Journal of School Psychology, 46(3), 281-314.

Alessi, G. (1988). Diagnosis diagnosed: A systemic reaction. Professional School Psychology, 3, 145-151.

Anderson, V. (1991). The neuropsychology of learning disabilities: Assessment, diagnosis, and treatment. Unpublished doctoral dissertation. University of Melbourne.

Apel, K., & Swank, L. K. (1999). Second chances: Improving decoding skills in the older student. Language, Speech & Hearing Services in Schools, 30, 231-243.

Australian Bureau of Statistics. (2008). Adult Literacy and Life Skills Survey (ALLS 2006-2007). Retrieved July 29, 2008, from http://www.abs.gov.au/AUSSTATS/abs@.nsf/Lookup/4102.0Chapter6102008

Australian Government House of Representatives Enquiry. (1993). The Literacy Challenge. Canberra: Australian Printing Office.

Badian, N. A. (1994). Preschool prediction: Orthographic and phonological skills, and reading. Annals of Dyslexia, 44, 3-25.

Beck, I. L., & McKeown, M. G. (1991). Social studies texts are hard to understand: Mediating some of the difficulties. Language Arts, 68, 482-490.

Beck, I. L., Perfetti, C. A., & McKeown, M. G. (1982). The effects of long-term vocabulary instruction on lexical access and reading comprehension. Journal of Educational Psychology, 74, 506-521.

Berninger, V. W., & Abbott, R. D. (1994). Redefining learning disabilities. Moving beyond aptitude-achievement discrepancies to failure to respond to validated treatment protocols. In G. Reid Lyon (Ed.), Frames of reference for the assessment of learning disabilities. New views on measurement issues (pp. 163-184). Baltimore, MD: Brooks Publishing.

Binder, C., Haughton, E., & Bateman, B. (2002). Fluency: Achieving true mastery in the learning process. Professional Papers in Special Education. VA: University of Virginia Curry School of Education. Retrieved 1/2/2003 from http://curry.edschool.virginia.edu/sped/projects/ose/papers/

Bowers, P. G. (1995). Tracing symbol naming speed's unique contributions to reading disabilities over time. Reading and Writing: An Inter-Disciplinary Journal, 7, 189-216.

Bowers, P. G., & Wolf, M. (1993). Theoretical links among naming speed, precise timing mechanisms and orthographic skill in dyslexia. Reading and Writing: An Interdisciplinary Journal, 5(1), 69-85.

Bowey, J. A., & Muller, D. (2005). Phonological recoding and rapid orthographic learning in third-grade children’s silent reading: a critical test of the self-teaching hypothesis, Journal of Experimental Child Psychology, 92, 203–219.

Bradley, L., & Bryant, P. (1983). Categorizing sounds and learning to read - A causal connection. Nature, 301, 419-421.

Brigance, A. H. (2000). Comprehensive Inventory of Basic Skills-Revised. Australia: Hawker Brownlow.

Brown, I. S. & Felton, R. H. (1990). Effects of instruction on beginning reading skills in children at risk for reading disability. Reading & Writing: An Interdisciplinary Journal, 2, 223-241.

Bruck, M. (1992). Persistence of dyslexics' phonological awareness deficits. Developmental Psychology, 28, 874-886.

California Department of Education. (1999). Reading/language arts framework for California public schools: Kindergarten through Grade Twelve. Retrieved June 3, 2000. from http://www.cde.ca.gov/cdepress/lang_arts.pdf

Campton, D. L., & Carlisle, J. F. (1994). Speed of word recognition as a distinguishing characteristic of reading disabilities. Educational Psychology Review, 6, 115 – 140.

Carnine, D. (2003, Mar 13). IDEA: Focusing on improving results for children with disabilities. Hearing before the Subcommittee on Education Reform Committee on Education and the Workforce United States House of Representatives. Retrieved July 11, 2005, from http://edworkforce.house.gov/hearings/108th/edr/idea031303/carnine.htm

Castles, A., & Coltheart, M. (1993). Varieties of developmental dyslexia. Cognition 47, 14-180.

Castles, A., & Coltheart, M. (2004). Is there a causal link from phonological awareness to success in learning to read? Cognition, 91, 77-111.

Chall, J. S. (1967). The great debate. New York: McGraw Hill.

Collier, K. (2008, October 18). The ABC of ignorance. Herald Sun, p.9.

Cornwall, A. (1992). The relationship of phonological awareness, rapid naming and verbal memory to severe reading and .disability. Journal of Learning Disabilities, 25, 532-538.

Coyne, M. D., Kame'enui, E. J., & Simmons, D. C. (2004). Improving beginning reading instruction and intervention for students with LD: Reconciling "All" with "Each". Journal of Learning Disabilities, 37(3), 231-239.

Cronin, J., Dahlin, M., Adkins, D., & Gage Kingsbury, G. (2007). The proficiency illusion. Thomas B. Fordham Foundation. Retrieved October 1, 2007, from http://edexcellence.net/template/page.cfm?id=264.

Crowley, K., Shrager, J. & Siegler, R. S. (1997). Strategy discovery as a competitive negotiation between metacognitive and associative mechanisms. Developmental Review, 17, 462-489.

Cuttance, P. (1998). Quality assurance reviews as a catalyst for school improvement in Australia. In A. Hargreaves, A. Lieberman, M. Fullan., & D. Hopkins (Eds.), International handbook of educational change, Part II (pp. 1135-1162). Dordrecht: Kluwer Publishers.

Denckla, M., & Rudel, R. G. (1974). Rapid "automatized" naming of pictured objects, colours, letters and numbers by normal children, Cortex, 10, 186-202.

Department for Education and Employment. (1998). The National Literacy Strategy: Framework for Teaching. London: Crown.

Department of Education, Employment & Training. (2001). Consistency Project: Links between curriculum frameworks in Vic, SA and QLD in English. Retrieved November 11, 2004, from http://www.sofweb.vic.edu.au/assess/consist/englink.htm

Department of Education, Science and Training (DEST). (2007). Parents’ attitudes to schooling. Canberra: Australian Government. Retrieved, 16/10/2008, from http://www.dest.gov.au/NR/rdonlyres/311AA3E6-412E-4FA4-AC01-541F37070529/16736/ParentsAttitudestoSchoolingreporMay073.rtf

Dixon, R., Engelmann, S. Bauer, M.M., Steely, D., & Wells, T. (1998). Spelling Mastery. Chicago: Science Research Associates.

Dunn, L. M., & Dunn, L. M. (1997). Peabody Picture Vocabulary Test. Circle Pines, MN: American Guidance Service.

Eden, G. F., Stein, J. F., Wood, H. M., & Wood, F. B. (1995a). Temporal and spatial processing in reading disabled and normal children. Cortex, 31, 451-468.

Eden, G. F., Stein, J. F., Wood, H. M., & Wood, F. B. (1995b). Verbal and visual problems in reading disability. Journal of Learning Disabilities, 28(5), 272-290.

Ehri, L. C. (1998). Presidential address. In Joanna P. Williams (Ed.), Scientific Studies of Reading Vol. 2, (No. 2, pp. 97-114). Mahwah, NJ: Lawrence Erlbaum.

Elbro, C., Nielsen, I., & Petersen, D. K. (1994). Dyslexia in adults: Evidence for deficits in non-word reading and in the phonological representation of lexical items. Annals of Dyslexia, 44, 205-226.

Engelmann, S. & Bruner, E. C. (1988). Reading Mastery. Chicago, Ill: Science Research Associates.

Engelmann, S. Haddox, P., & Bruner, E. (1983). Teach Your Child to Read in 100 Easy Lessons. New York: Simon & Schuster.

Engelmann, S., & Osborn, J. (1999). Language for Learning. Columbus, OH: SRA/McGraw-Hill.

Engelmann, S., & Silbert, J. (1983). Expressive Writing I. Desoto, TX: SRA/McGraw-Hill.

Engelmann, S., & Silbert, J. (1991). Reasoning and Writing. Desoto, TX: SRA/McGraw-Hill.

Engelmann, S., Haddox, P., Hanner, S., & Osborne, J. (1989). Corrective Reading: Comprehension. Chicago: Science Research Associates.

Engelmann, S., Meyer, L., Carnine, L., & Johnson, G. (1999). Corrective Reading: Decoding. Chicago: Science Research Associates.

Fehring, H. (2001). Literacy assessment and reporting: Changing practices. 12th European Conference on Reading, RAI Dublin, 1st - 4th July. Retrieved November 1, 2006, from http://sece.eu.rmit.edu.au/staff/fehring/irish.htm

Felton, R. H. (1992). Early identification of children at risk for reading disabilities. Topics in Early Childhood Special Education, 12, 212-229.

Felton, R. H. (1993). Effects of instruction on the decoding skills of children with phonological-processing problems. Journal of Learning Disabilities, 26, 583-589.

Foorman, B. R. (1995). Research on "the great debate" code-oriented versus whole language approaches to reading instruction. School Psychology Review, 24, 376-392.

Fuchs, L. S., & Fuchs, D. (1992). Identifying a measure for monitoring student reading progress. School Psychology Review, 21, 45-58.

Fuchs, L. S., Fuchs, D., Hosp, M. K., and Jenkins, J. R. (2001). Oral reading fluency as an indicator of reading competence: A theoretical, empirical, and historical analysis. Scientific Studies of Reading, 5, 239-256.

Gaillard, R., Naccache, L., Pinel, P., Clemenceau, S., Volle, E., Hasboun, D., Dupont, S., Baulac, M., Dehaene, S., Adam, C., Cohen, L. (2006). Direct intracranial, fMRI and lesion evidence for the causal role of left inferotemporal cortex in reading. Neuron, 50, 191-204.

Gillam, R. B., & Van Kleek, A. (1996). Phonological awareness training and short-term working memory: Clinical implication. Topics in Language Disorders, 17(1), 72-81.

Good, R. H., & Kaminski, R. A. (2002). Dynamic indicators of basic early literacy skills (6th ed.). Eugene, OR: Institute for Development of Educational Achievement.

Goodman, K. S. (1986). What's whole in whole language. Richmond Hill, Ontario: Scholastic.

Greenberg, D., Ehri, L. C., & Perin, D. (1997). Are word reading processes the same or different in adult literacy students and third-fifth graders matched for reading level? Journal of Educational Psychology, 89, 262-275.

Gresham, F. (2001). Responsiveness to intervention: An alternative approach to the identification of learning disabilities. In R. Bradley, L. Danielson, & D. P. Hllahan (Eds.), Identification of learning disabilities: Research in practice (pp.467–519). Mahwah, NJ: Erlbaum.

Grossen, B. (1997). A synthesis of research on reading from the National Institute of Child Health and Human Development. Retrieved September 1, 2008, from http://www.nrrf.org/synthesis_research.htm.

Hart, B., & Risley, T. R. (1995). Meaningful differences in the everyday experiences of young American children. Baltimore, MD: Paul H. Brookes.

Hart, B., & Risley, T. R. (2003, Spring). The early catastrophe: The 30 million word gap. American Educator. Retrieved April 11, 2003 from http://www.aft.org/american_educator/spring2003/catastrophe.html

Hatcher, P. J. (1994). Sound Linkage: An integrated programme for overcoming reading difficulties. London: Whurr Publishers.

Hattie, J. A., Clinton, J., Thompson, M., & Schmidt-Davies, H. (1995). Identifying highly accomplished teachers: A validation study. Greensboro, NC: Center for Educational Research and Evaluation, University of North Carolina.

Hempenstall, K. (1995). The Picture Naming Test. Unpublished manuscript. Royal Melbourne Institute of Technology.

Hempenstall, K. (1996). The gulf between educational research and policy: The example of direct instruction and whole language. Behaviour Change, 13, 33-46.

Hempenstall, K. (1998). Miscue analysis: A critique. Australian Journal of Learning Disabilities, 3(4), 32-37.

Hempenstall, K. (2001). School-based reading assessment: Looking for vital signs. Australian Journal of Learning Disabilities, 6, 26-35.

Hempenstall, K. (2003). The three-cueing system: Trojan horse? Australian Journal of Learning Disabilities, 8(3), 15-23.

Hempenstall, K. (2005). The Whole Language-Phonics controversy: An historical perspective. Australian Journal of Learning Disabilities, 10(3 & 4), 19-33.

Hempenstall, K. (2006). What does evidence-based practice in education mean? Australian Journal of Learning Disabilities, 11(2), 83-92.

Henty, A. (1993). Speech Pathology Phonological Screening Awareness Test. Unpublished manuscript.

Hill, P. (1995). School effectiveness and improvement: Present realities and future possibilities. Dean's Lecture: Paper presented at Melbourne University, May 24. Retrieved September 2, 2000, from http://www.edfac.unimelb.edu.au/Seminars/dean_lec/list.html

Hill, P. (2000, August 8). Our desperate schools. The Age, p.11.

Hirsch, E. D. (2006, Spring).What do reading comprehension tests measure? Knowledge. American Educator, 30(1). Retrieved October 21, 2007, from www.aft.org/pubsreports/american_educator/issues/spring06/tests.htm

Hoien, T., Lundberg, I., Stanovich, K. E., & Bjaalid, I-K. (1995). Components of phonological awareness. Reading and Writing: An Inter-Disciplinary Journal, 7, 171-188.

Hoover, W. & Gough, P. (1990). The simple view of reading. Reading and Writing: An Inter-Disciplinary Journal, 2, 127-160.

Howell, K. W., & Nolet, V. (2000). Curriculum-based evaluation: Teaching and decision making (3rd ed.). Belmont, CA: Wadsworth.

Hurford, D. P., Darrow, L. J., Edwards, T. L., Howerton, C. J., Mote, C. R., Schauf, J. D., & Coffey, P. (1993). An examination of phonemic processing abilities in children during their first grade year. Journal of Learning Disabilities, 26, 167-177.

Hurford, D. P., Schauf, J. D., Bunce, L., Blaich, T., & Moore, K. (1994). Early identification of children at risk for reading disabilities. Journal of Learning Disabilities, 27, 371-382.

Joseph, J., Noble, K., & Eden, G. (2001). The neurobiological basis of reading. Journal of Learning Disabilities, 34, 566-579.

Joshi, M., & Aaron, P. (2002). Naming speed and word familiarity as confounding factors in decoding. Journal of Research in Reading, 25(2), 160 –171.

Juel, C. (1988). Learning to read & write: A longitudinal study of 54 children from first through fourth grades. Journal of Educational Psychology, 80, 437-447.

Kame'enui, E. J., Simmons, D. C., & Coyne, M. D. (2000). Schools as host environments: Toward a schoolwide reading improvement model. Annals of Dyslexia, 50, 33-51.

Kaplan, E., Goodlass, H., & Weintraub, S. (1983). Boston Naming Test. Philadelphia, PA: Lea and Febiger.

Koretz, D. (1992). NAEP and national testing: Issues and implications for educators. NASSP Bulletin, 76(545), 30-40. Washington, DC: Rand Corp.

Laberge, D., & Samuels, S. (1974). Toward a theory of automatic information processing in reading. Cognitive Psychology, 6, 293-323.

Labov, L. (2003). When ordinary children fail to read. Reading Research Quarterly, 38, 128-131.

Landi, N., Perfetti, C. A., Bolger, D. G., Dunlap, S. & Foorman, B. R. (2006). The role of discourse context in developing word form representations: A paradoxical relation between reading and learning. Journal of Experimental Child Psychology, 94(2), 114-133.

Leigh, A., & Ryan, C. (2008). How has school productivity changed in Australia? The Australian National University, Canberra. Retrieved September 12, 2008, from http://econrsss.anu.edu.au/~aleigh/pdf/SchoolProductivity.pdf

Levin, I., Shatil-Carmon, S., & Asif-Rave, O. (2006). Learning of letter names and sounds and their contribution to word recognition. Journal of Experimental Child Psychology, 93(2), 139-165.

Lewkowicz, N. K. (1980). Phonemic awareness training: What to teach and how to teach it. Journal of Educational Psychology, 72, 686-700.

Lindamood, C. H., & Lindamood, P. C. (1979). Lindamood Auditory Conceptualization Test. Allen, TX: DLM Teaching Resources.

Livingstone, T. (2006, 14 Oct). Rating the system. The Courier Mail, p.53.

Louden, W., Chan, L. K. S., Elkins, J., Greaves, D., House, H., Milton, M., Nichols, S., Rivalland, J., Rohl, M., & van Kraayenoord, C. (2000). Mapping the territory—Primary students with learning difficulties: Literacy and numeracy (Vols. 1-3). Canberra: Department of Education Training and Youth Affairs.

Love, E., & Reilly, S. (1995). A Sound Way: Phonological Awareness - Activities for Early Literacy. Melbourne: Longman Cheshire.

Lovett, M. W., Steinbach, K. A., & Frijters, J. C. (2000). Remediating the core deficits of developmental reading disability: A double-deficit perspective. Journal of Learning Disabilities, 33, 334-342.

Lyon, G. R. (1995). Toward a definition of dyslexia. Annals of Dyslexia, 45, 3-27.

Lyon, G. R. (1999, December, 12). Special education in state is failing on many fronts. Los Angeles Times, p. A1. Retrieved September 10, 2000, from http://www.latimes.com/news/state/reports/specialeduc/lat_special991212.htm

Lyon, G. R. (2001). Measuring success: Using assessments and accountability to raise student achievement. Subcommittee on Education Reform Committee on Education and the Workforce U.S. House of Representatives Washington, D.C. Retrieved 1/2/2003 from http://www.nrrf.org/lyon_statement3-01.htm

Manis, F. R. Doi, L. M., & Bhadha, B. (2000). Naming speed, phonological awareness, and orthographic knowledge in second graders. Journal of Learning Disabilities, 33, 325-333.

Marshall R. M. & Hynd, G. W. (1993). Neurological basis of learning disabilities. In William W. Bender (Ed.) Learning disabilities: Best practices for professionals. Stoneham, MA: Butterworth-Heinemann.

McBride-Chang, C. (1999). The ABCs of the ABCs: the development of letter-name and letter-sound knowledge. Merrill-Palmer Quarterly, 45, 285-296.

MCEETYA. (2008). National Assessment Program: Literacy and Numeracy. Ministerial Council on Education, Employment, Training, and Youth Affairs. Retrieved September 20, 2008, from http://www.naplan.edu.au/home_page.html

McGregor, K. K., & Leonard, L. B. (1995). Intervention for word-finding deficits in children. In M. Fey, J. Windsor, & S. Warren, (Eds.), Language intervention: Preschool through the elementary years, (pp. 85-105). Baltimore, MD: Paul H. Brookes.

Metsala, I. L., & Ehri, L. C. (1998). Word recognition in beginning literacy. Mahwah, NJ: Lawrence Erlbaum Associates.

Miller, J., & Schwanenflugel, P. J. (2006). Prosody of syntactically complex sentences in the oral reading of young children. Journal of Educational Psychology, 98(4), 839-853.

Miller, L. L., & Felton, R. H. (2001). "It's one of them ... I don't know": Case study of a student with phonological, rapid naming, and word-finding deficits. Journal of Special Education, 35, 125-133.

Nagy, W. E. (1998). Increasing students’ reading vocabularies. Presentation at the Commissioner’s Reading Day Conference, Austin, Texas.

Nagy, W. E., & Anderson, R. C. (1984). How many words are there in printed English? Reading Research Quarterly, 19, 304-330.

Nation, K., Angell, P., & Castles, A. (2007). Orthographic learning via self-teaching in children learning to read English: Effects of exposure, durability, and context. Journal of Experimental Child Psychology, 96(1), 71-84.

National Inquiry into the Teaching of Literacy (2005). Teaching reading - A review of the evidence-based research literature on approaches to the teaching of literacy, particularly those that are effective in assisting students with reading difficulties. Australian Government: Department of Education, Science and Training. Retrieved November 1, 2007, from www.dest.gov.au/nitl/documents/literature_review.pdf.

National Institute of Child Health and Human Development (2000). National Reading Panel: Teaching children to read. Retrieved September 25, 2002, from http://www.nationalreadingpanel.org.

Neale, M. D. (1988). Neale Analysis of Reading Ability-Revised. Melbourne, Australia: ACER.

Neilson, R. (2003). Sutherland Phonological Awareness Test-Revised. NSW, Australia: Language Speech & Literacy Services.

Nicholson, T., & Whyte, B. (1992). Matthew effects in learning new words while listening to stories. In C. K. Kinzer & D. J. Leu (Eds.), Literacy research, theory, and practice: Views from many perspectives: Forty-first Yearbook of the National Reading Conference (pp. 499-503). Chicago, IL: The National Reading Conference.

Oakhill, J. V., & Garnham, A. (1988). Becoming a skilled reader. Oxford: Basil. Blackwell.

O'Connor, R. E., Bell, K. M., Harty, K. R., Larkin, L. K., Sackor, S. M., & Zigmond, N. (2002). Teaching reading to poor readers in the intermediate grades: A comparison of text difficulty. Journal of Educational Psychology, 94, 474-485.

OECD. (2004). Adults at low literacy level (most recent) by country. International Adult Literacy Survey (IALS). Retrieved, 16/10/2008, from http://www.nationmaster.com/graph/edu_lit_adu_at_low_lit_lev-education-literacy-adults-low-level

Office of the Victorian Auditor General. (2003). Improving literacy standards in government schools. Retrieved 10/10/2004 from http://www.audit.vic.gov.au/reports_par/Literacy_Report.pdf

Paton, G. (2008, 25 Oct). GCSE standards 'deliberately lowered' to make sure pupils pass. Telegraph.co.uk. Retrieved October 25, 2008, from http://www.telegraph.co.uk/news/newstopics/politics/education/3254233/GCSE-standards-deliberately-lowered-to-make-sure-pupils-pass.html

Pearson, P. D., & Hamm, D. N. (2005). The assessment of reading comprehension: A review of practices – past, present, and future (pp. 13-69). In S. G. Paris, & S. A. Stahl (Eds.), Children's reading comprehension and assessment. (pp. 131-160). Mahwah, NJ: Lawrence Erlbaum Associates Publishers.

Pearson, P. D., Hiebert, E. H., & Kamil, M. L. (2007). Vocabulary assessment: What we know and what we need to know. Reading Research Quarterly, 42(2), 282-296.

Perfetti, C. A. (1985). Reading ability. Oxford, UK: Oxford University Press.

Perfetti, C. A. (1992). Introduction. In N. N. Singh & I. L. Beale (Eds), Learning disabilities: Nature, theory and treatment (pp.1-22). New York: Springer-Verlag.

Perfetti, C. A., & Hogaboam, T. (1975). Relationship between single word decoding and reading comprehension skill. Journal of Educational Psychology, 67, 461-469.

Perfetti, C. A., & Lesgold, A. M. (1977). Coding and comprehension in skilled reading and implications for reading instruction. In L. B Resnick & P. A Weaver (Eds.), Theory and practice of early reading (Vol. 1 pp 57-84) Hillsdale, NJ: Earlbaum.

Pressley, M. (2000). What should comprehension instruction be the instruction of? In M. L. Kamil, P. B. Mosenthal, P. D. Pearson, & R. Barr (Eds.), Handbook of reading research (vol. 3, pp. 545-561). Mahwah, NJ: Erlbaum.

Pressley, M. (2001, September). Comprehension instruction: What makes sense now, what might make sense soon. Reading Online, 5(2). Retrieved 2 October, 2003, from http://www.readingonline.org/articles/art_index.asp?HREF=/articles/handbook/pressley/index.html

Primary National Strategy (2006). Primary framework for literacy and mathematics. UK: Department of Education and Skills. Retrieved 26 October, 2006, from http://www.standards.dfes.gov.uk/primaryframeworks/

Pugh, K. P., Mencl, W.E., Jenner, A. R., Katz, L., Frost, S. J., Lee, J. R., Shaywitz, S. E., & Shaywitz, B. A. (2002). Neuroimaging studies of reading development and reading disability. Learning Disabilities Research & Practice, 16, 240-249.

Rack, J. P., Snowling, M., & Olson, R. K. (1992). The nonword reading deficit in developmental dyslexia: A review. Reading Research Quarterly, 27, 39-53.

Rathvon, N. (2004). Early reading assessment: A practitioner’s handbook. New York: Guilford Press.

Rayner, K., Foorman, B. R., Perfetti, C. A., Pesetsky, D., & Seidenberg, M. S. (2001). How psychological science informs the teaching of reading. Psychological Science in the Public Interest, 2, 31-74. Retrieved February 1, 2003, from www.psychologicalscience.org/newsresearch/publications/journals/pspi2_2.html

Richards, T. L., Aylward, E. H., Berninger, V. B., Field, K. M., Grimme, A. C., Richards, A. L., & Nagy, W. (2006). Individual fMRI activation in orthographic mapping and morpheme mapping after orthographic or morphological spelling treatment in child dyslexics. Journal of Neurolinguistics, 19(1), 56-86.

Richards, T. L., Corina, D., Serafini, S., Steury, K., Echelard, D. R., Dager, S. R., Marro, K., Abbott, R. D., Maravilla, K. R., & Berninger, V. W. (2000). The effects of a phonologically-driven treatment for dyslexia on lactate levels as measured by Proton MRSI. American Journal of Neuroradiology, 21, 916-922. Retrieved 12/2/03 from http://faculty.washington.edu/toddr/dyslexic2.htm

Richards, T. L., Dager, S. R., Corina, D., Serafini, S., Heide, A. C., Steury, K., Strauss, W., Hayes, C. E., Abbott, R. D., Craft, S., Shaw, D., Posse, S., & Berninger, V. W. (1999). Dyslexic children have abnormal brain lactate response to reading-related language tasks. American Journal of Neuroradiology, 20, 1393-1398.

Roehrig, A. D., Petscher, Y., Nettles, S. M., Hudson, R. F., & Torgesen, J. T. (2008). Accuracy of the DIBELS Oral Reading Fluency measure for predicting third grade reading comprehension outcomes. Journal of School Psychology, 46(3), 343–366.

Roid, G. H. (2003). The Stanford-Binet Intelligence Scale: Fifth Edition. Itasca, Ill: Riverside Publishing Company.

Rose, J. (2006). Independent review of the teaching of early reading. Bristol: Department for Education and Skills. Retrieved April 12, 2006, from www.standards.dfes.gov.uk/rosereview/report.pdf

Rosner, J. (1975). Helping children to overcome learning disabilities. Navato, CA: Academic Therapy.

Rubin, H., Rotella, T., Schwartz, L., & Bernstein, S. (1991). The effect of phonological analysis training on naming performance. Reading & Writing: An Interdisciplinary Journal, 3, 1-10.

Satz, P., Fletcher, J. M., Clark, W., & Morris, R. (1981). Lag, deficit, rate and delay constructs in specific learning disabilities: A re-examination. In A. Ansara, N. Geschwind, A. Galaburda, M. Albert, & N. Gartrell (Eds.), Sex differences in dyslexia (pp. 129-150). Towson, MD: The Orton Dyslexia Society.

Savage, R., & Frederickson, N. (2005). Evidence of a highly specific relationship between rapid automatic naming of digits and text-reading speed. Brain and Language, 93, 152–159.

Savage, R.S., & Frederickson, N. (2006). Beyond phonology: What else is needed to describe the problems of below-average readers and spellers? Journal of Learning Disabilities, 39(5), 399-413.

Scammacca, N., Roberts, G., Vaughn. S., Edmonds, M., Wexler, J., Reutebuch, C. K., & Torgesen, J. K. (2007). Interventions for adolescent struggling readers: A meta-analysis with implications for practice. Portsmouth, NH: RMC Research Corporation, Center on Instruction.

Scarborough, H. (2003). Connecting early language and literacy to later reading (dis)abilities: Evidence, theory, and practice. In S. Newman & D. Dickinson (Eds.), Handbook of early literacy research (pp. 97–110). New York: The Guilford Press.

Scarborough, H. S. (1998). Early identification of children at risk for reading disabilities. In B. K. Shapiro, P. J. Accardo, & A. J. Capute (Eds.), Specific reading disability (pp. 75-119). Timonium, MD: York Press.

Schatschneider, C., Francis, D. J., Foorman, B. R., Fletcher, J. M., & Mehta, P. (1999). The dimensionality of phonological awareness: An application of item response theory. Journal of Educational Psychology, 91(3), 439-449.

Shankweiler, D., Lundquist, E., Dreyer, L. G., & Dickinson, C. C. (1996). Reading and spelling difficulties in high school students: Causes and consequences. Reading and Writing: An Inter-Disciplinary Journal, 8, 267-294.

Shankweiler, D., Lundquist, E., Katz, L., Stuebing, K. K., Fletcher, J. M., Brady, S., Fowler, A., Dreyer, L. G., Marchione, K. E., Shaywitz, S. E., & Shaywitz, B. A. (1999). Comprehension and decoding: Patterns of association in children with reading difficulties. Scientific Studies of Reading, 3, 69-94.

Share, D. L. (1995). Phonological recoding and self-teaching: Sine qua non of reading acquisition. Cognition, 55, 151-218.

Share, D. L. (1999). Phonological recoding and orthographic learning: A direct test of the self-teaching hypothesis. Journal of Experimental Child Psychology, 72(2), 95-129.

Share, D. L. (2004). Orthographic learning at a glance: On the time course and developmental onset of self-teaching. Journal of Experimental Child Psychology 87(4), 267-298.

Share, D. L., & Stanovich, K. E. (1995). Cognitive processes in early reading development: accommodating individual differences into a model of acquisition. Issues in Education, 1, 1-57.

Shaywitz, B. A., Shaywitz, S. E., Blachman, B. A., Pugh K. R., Fulbright, R. K., Skudlarski, P., Mencl, W. E., Constable, R. T., Holahan, J. M., Marchione, K. E., Fletcher, J. M., Lyon, G. R., & Gore, J. C. (2004). Development of left occipitotemporal systems for skilled reading in children after a phonologically- based intervention. Biological Psychiatry, 55, 926-33.

Shaywitz, S. (2003). On the mind of a child: A conversation with Sally Shaywitz. Educational Leadership, 60(7), 6-10.

Shaywitz, S. E., Fletcher, J. M., Holahan, J. M., Shneider, A. E., Marchione, K. E., Stuebing, K. K., Francis, D. J., Pugh, K. R., & Shaywitz, B. A. (1999). Persistence of dyslexia: The Connecticut longitudinal study at adolescence. Pediatrics, 104, 1351-1339.

Shinn, M. R., Good, R. H., Knutson, N., Tilly, W. D., & Collins, V. (1992). Curriculum-based measurement of oral reading fluency: A confirmatory analysis of its relation to reading. School Psychology Review, 21, 459-479.

Siegel, L. S. (1993). The development of reading. Advances in Child Development and Behaviour, 24, 63-97.

Simos, P. G., Fletcher, J. M., Sarkari, S., Billingsley-Marshall, R. L., Denton, C. A., Papanicolaou, A. C. (2007). Intensive instruction affects brain magnetic activity associated with reading fluency in children with dyslexia. Journal of Learning Disabilities, 40(1), 37-48.

Simos, P., Fletcher, J., Bergman, E., Breier, J., Foorman, B., Castillo, E., Davis, R., Fitzgerald, M., & Papanicolaou, A. (2002). Dyslexia-specific brain activation profile becomes normal following successful remedial training. Neurology, 58, 1203-1212.

Snow, C. E., Burns, S., & Griffin, P. (Eds.) (1998). Preventing reading difficulties in young children. Report of the National Research Council. Retrieved September 2, 1999, from http://www.nap.edu/readingroom/books/reading/.

Spadafore, G. J. (1983). Spadafore Diagnostic Reading Test. Novato, CA: Academic Therapy Publications.

Sparks, R. L. (1995). Phonemic awareness in hyperlexic children. Reading and Writing: An Inter-Disciplinary Journal, 7, 217-235.

Spear-Swerling, L. (1998, August). The use and misuse of processing tests. LD In Depth. Retrieved September 20, 1998, from ldonline.org/ld_indepth/assessment/swerling_assessment.html

Spector, J. (1992). Predicting progress in beginning reading: Dynamic assessment of phonemic awareness. Journal of Educational Psychology, 84, 353-363.

Speece, D. L., Mills, C., Ritchey, K. D., & Hillman, E. (2003). Initial evidence that letter fluency tasks are valid indicators of early reading skill. Journal of Special Education, 36, 223-233.

Stage, S. A., Sheppard, J., Davidson, M. M., & Browning, M. M. (2001). Prediction of first-graders' growth in oral reading fluency using kindergarten letter fluency. Journal of School Psychology, 9(3), 225-237.

Stanovich, K. E. (1986). Matthew effects in reading: Some consequences of individual differences in the acquisition of literacy. Reading Research Quarterly, 21, 360-406.

Stanovich, K. E. (1988a). The right & wrong places to look for the cognitive locus of reading disability. Annals of Dyslexia, 38, 154-157.

Stanovich, K. E. (1988b). Explaining the differences between the dyslexic and the garden-variety poor reader: The phonological-core variable-difference model. Journal of Learning Disabilities, 21, 590-612.

Stanovich, K. E. (1991). Discrepancy definitions of reading disability: Has intelligence led us astray? Reading Research Quarterly, 26(1), 7-29.

Stanovich, K. E. (1992). Speculation on the causes and consequences of individual differences in early reading acquisition. In Phillip P. Gough, Linnea C. Ehri, & Rebecca Treiman (Eds.), Reading acquisition. (pp.307-341). Mahwah, NJ: Lawrence Erlbaum.

Stanovich, K. E. (1993). Does reading make you smarter? Literacy and the development of intelligence. Advances in Child Development and Behavior, 24, 133-179.

Stanovich, K. E. (2000). Progress in understanding reading: Scientific foundations and new frontiers. New York: Guilford Press.

Strutt, J. (2007). Students fail on 'three Rs' test. The West Australian, Monday 10 December. Retrieved September 20, 2008, from http://www.platowa.com/Breaking_News/2007/2007_12_10.html

Stuart, M. (1995). Prediction and qualitative assessment of five and six-year-old children's reading: A longitudinal study. British Journal of Educational Psychology, 65, 287-296.

Swanson, H. L. (2001a). Research on interventions for adolescents with learning disabilities: A meta-analysis of outcomes related to higher-order processing. The Elementary School Journal, 101, 331-348.

Swanson, H. L. (2001b). Searching for the best model for instructing students with learning disabilities.  Focus on Exceptional Children, 34(2), 1-14.

Swanson, L. B. (1989). Analysing naming speed-reading relationships in children. Unpublished doctoral dissertation, University of Waterloo, Ontario.

Swingley, D. (2008). The roots of the early vocabulary in infants' learning from speech. Current Directions in Psychological Science, 17(5), 308-312.

The Psychological Corporation (2007). AIMSweb progress monitoring and response to intervention system. San Antonio, TX: Pearson. Retrieved October 22, 2008, from www.aimsweb.com

Torgesen, J. K. (1998, Spring/Summer). Catch them before they fall: Identification and assessment to prevent reading failure in young children. American Educator. Retrieved October 2, 1999, from http://www.ldonline.org/ld_indepth/reading/torgeson_catchthem.html

Torgesen, J. K. (2000). Individual differences in response to early interventions in reading: The lingering problem of treatment resistors. Learning Disabilities Research and Practice, 15, 55– 64.

Torgesen, J. K., & Bryant, B. R. (1994). Test of Phonological Awareness. Austin, TX: Pro-Ed.

Torgesen, J. K., Wagner, R. J., & Rashotte, C. A. (1994). Longitudinal studies of phonological processing & reading. Journal of Learning Disabilities, 27, 276-286.

Torgesen, J. K., Wagner, R. J., & Rashotte, C. A. (1999). Test of Word Reading Efficiency (TOWRE). Austin, TX: PRO-ED Inc.

Torgesen, J. K., Wagner, R., Rashotte, C., Alexander, A., & Conway, T. (1997). Preventative and remedial interventions for children with severe reading disabilities. Learning Disabilities: A Multidisciplinary Journal, 8, 51-61.

Turkeltaub, P. E., Flowers, D. L., Verbalis, A., Miranda, M., Gareau, L., & Eden, G. F. (2004). The neural basis of hyperlexic reading: An fMRI case study. Neuron, 41, 1-20.

Tzuriel, D. (2000). Dynamic assessment of young children: Educational and intervention perspectives Educational Psychology Review, 12, 385-435.

UNICEF. (2002). A league table of educational disadvantage in rich nations. Innocenti Report Card No.4, November 2002. Florence, Italy: Innocenti Research Centre.

Vellutino, F. R., & Scanlon, D. M. (1987). Phonological coding, phonological awareness and reading ability: Evidence from a longitudinal & experimental study. Merrill-Palmer Quarterly, 33, 321-363.

Vellutino, F. R., Scanlon, D. M., Sipay, E. R., Small, S. G., Pratt, A., Chen, R., & Denckla, M. B. (1996). Cognitive profiles of difficult to remediate and readily remediated poor readers: Early intervention as a vehicle for distinguishing between cognitive and experiential deficits as basic causes of specific reading disability. Journal of Educational Psychology, 88, 601-638.

Vervaeke, S.-L., McNamara, J. K., & Scissons, M. (2007, April). Kindergarten screening for reading disabilities. Journal of Applied Research on Learning, 1(1), 1-19.

Wade, B., & Moore, M. (1993). Experiencing special education. Buckingham: Open University Press.

Wagner, R. K., & Torgesen, J. K. (1987). The nature of phonological processing and its causal role in the acquisition of reading skills. Psychological Bulletin, 101, 192-212.

Wagner, R. K., Torgesen, J. K., & Rashotte, C. A. (1999). Comprehensive Test of Phonological Processing. Austin, TX: Pro-Ed.

Weaver, C. (1988). Reading process & practice: From socio-psycholinguistics to whole language. Portsmouth, NH: Heinemann.

Wechsler, D. (1997). Wechsler Adult Intelligence Scale (3rd ed.). Austin, TX: The Psychological Corporation.

Wechsler, D. (2001). Wechsler Individual Achievement Test-II. San Antonio, TX: Harcourt Brace.

Wechsler, D. (2002). The Wechsler Preschool and Primary Scales of Intelligence-Third Edition: WPPSI-III. San Antonio, TX: The Psychological Corporation.

Wechsler, D. (2003). Wechsler Intelligence Scale for Children – 4th Edition (Australian Adaptation). Australia: Harcourt Brace.

Wiederholt, J. L., & Bryant, B. R. (2001). Gray Oral Reading Tests-4th Edition. Austin, TX: Pro-Ed.

Wiig, E. H., Zureich, P., & Chan, H. H. (2000). A clinical rationale for assessing rapid automatized naming in children with language disorders. Journal of Learning Disabilities, 33, 359-371.

Wolf, M., & Bowers, P. G. (1999). The "DoubleDeficit Hypothesis" for the developmental dyslexias. Journal of Educational Psychology, 91, 1-24.

Wolf, M., & Bowers, P. G. (2000). Naming-speed processes and developmental reading disabilities: An introduction to the special issue on the double-deficit hypothesis. Journal of Learning Disabilities, 33, 322-341.

Wolf, M., & Katzir-Cohen, T. (2001). Reading fluency and its intervention. Scientific Studies of Reading. (Special Issue on Fluency. Editors: E. Kameenui & D. Simmons), 5, 211-238.

Wolf, M., Miller, L., & Donnelly, K. (2000). Retrieval, automaticity, vocabulary elaboration, orthography (RAVE-O): A comprehensive, fluency-based reading intervention program. Journal of Learning Disabilities, 33, 375-386.

Wood, F. B., & Felton, R. H. (1994). Separate linguistic and attentional factors in the development of reading. Topics in Language Disorders, 14(4), 42-57.

Woodcock, R. W. (1998). Woodcock Reading Mastery Test – Revised NU. Circle Pines, MN: American Guidance Service.

Yopp, H. K. (1995). A test for assessing phonemic awareness in young children. The Reading Teacher, 49(1), 20-29.

 

Figure 1: Assessment Results: Possible Interpretation

Client

Test Results

Interpretation/Possible Causes

Interpretation might have been ruled out by..

Interventions/Recommendations

 

 

“Donna”

diagnosis

Dyslexia

WISC-IV:

o   Average or ↑ FSIQ

o   May or may not demonstrate a discrepancy between VCI and PRI.

o   Maybe ↓WMI or ↓PSI

Brigance:

o   ↑LC ↓RC

CTOPP:

o   ↓↓ score on any or all subtests

DIBELS:

o   ↓ fluency – at risk

TOWRE:

o   ↓↓ scores on PDE and probably on ↓ or ↓↓ on SWE

SAST:

o   ↓↓ critically low

TOWL

o   ↓ or ↓↓

PPVT-3:

o   ↑↑ or ↑

CRP Placement Test:

o   ↓↓ Decoding

o   Comprehension perhaps

o   Level of program will vary depending on age and ability level

Dyslexia: has been characterised by the following:

o   A score on a standardised test in reading is significantly (1, 1.5, or 2 standard deviations) below IQ score, indicating a learning difficulty which is intrinsic to the individual and due to a central nervous system dysfunction.

o   A significant discrepancy between ↑LC and ↓RC

o   Low score on a standardised test of phonological processing

o   Resistance to recognised reading intervention

Environmental Factors:

o   Inadequate instructional management

o   Failure of early identification and intervention

o   Poor or no initial teaching of phonological awareness, along with letters and their associated sounds.

Physical Factors:

o   Genetic predisposition – tendency to experience a phonological reading difficulty possibly inherited. 

o   History of ear infections?

o   Low FSIQ on WISC-IV, therefore no discrepancy between cognitive ability and academic functioning.

o   No discrepancy between RC and LC.

o   High or average score on tests of phonological processing

o   Vision and hearing difficulties.

o   Anxiety or attentional difficulties, behavioural or fatigue problems during assessment.

o   Corrective Reading (CRP): Decoding program or 100 Lessons (choose appropriate program/level according to age and reading level of client).

o   Given the level of difficulty experienced, emphasise the program’s repeat tasks until firm instruction.

o   If the client also meets criteria for the CRP: Comprehension program, the decoding program should be completed first, as it is likely that comprehension will improve as decoding skills improve.

o   Recommend to re-administer comprehension placement test following completion of each decoding program.

o   May need to boost intervention with a spelling program, e.g., Spelling Mastery Series, and a writing program, e.g., Expressive Writing.

o   Adapt home and school environment to accommodate for reading deficits: e.g.,

o   New information should be presented in a modality appropriate to the task.

o   Special consideration may need to be given e.g., modified exam and assignment situations.

o   Make provisions for key course texts to be taped and kept in the school library; use of digital voice recorder in class.

RC = Reading Comprehension; LC = Listening Comprehension; SWE = Sight Word Efficiency; PDE = Phonemic Decoding Efficiency; SAST = South Australian Spelling Test; WISC-IV = Wechsler Intelligence Scale for Children – Fourth Edition; FSIQ = Full Scale IQ; VCI = Verbal Comprehension Index; PRI = Perceptual Reasoning Index; WMI = Working Memory Index; PSI = Processing Speed Index; PPVT-3 = Peabody Picture Vocabulary Test; TOWL: Test of Written Language; CRP = Corrective Reading program


*************************************************************************************

“National minimum standards for literacy and numeracy are defined for assess ments in Reading, Writing and Numeracy at each year level. The national minimum standard for each year level is defined and located on a common underlying NAPLAN scale. Students achieving at the minimum standard have typically demonstrated only the basic elements of literacy and numeracy for their year level (p. 6) … The use of NAPLAN to make decisions about education provision between the states and federal government also delimits what is construed as valid and valuable education.”

Hardy, I. (2025). Australian schooling policy: A risky proposition. The Australian Educational The Australian Educational Researcher (2025) 52:209–230 https://doi.org/10.1007/s13384-024-00711-6

__________________________________________________________________

“It (re)examines data from an ethno-case study into the datafication of assessment and learning over one school year, in primary and secondary schooling contexts, to understand the uses of the National Assessment Program – Literacy and Numeracy (NAPLAN) in a new, critical light. We explore schools’ contemporary uses of NAPLAN – intended or otherwise – to demonstrate how the policy effects of NAPLAN have become insidiously submerged within the daily practices in schools. Drawing on interviews with 27 teachers and seven school leaders, classroom and staff meeting observations, and artefact data, we reveal the invisible yet profoundly altering presence of NAPLAN and its consequences. Specifically, we analyse the ways in which NAPLAN practices, structures and technologies are both hidden and yet manifestly altering as a) practices that disappear into their uses, becoming unidentifiable and routine; and b) practices that follow well-used pathways that further embed particular uses. We counter rhetoric of NAPLAN normativity and complacency, instead demonstrating that its current uses, while not originally intended, are insidious and profound.”

Daliri-Ngametua, R., Wescott, S., & McKay, A. (2023). On the uses and use of NAPLAN: the hidden effects of test-based data-centric accountabilities. Journal of Education Policy39(6), 843–860. https://doi.org/10.1080/02680939.2023.2273499

_____________________________________________________________________

“This study examines public engages, Sara Ahmed’s theorising engages, Sara Ahmed’s theorising opinion about the National Assessment Program – Literacy and Numeracy (NAPLAN) in Australia, based on an online survey conducted with a sample of New South Wales (NSW) residents (n = 2,017). Our study participants did recognize the benefits of NAPLAN despite their opinion generally trending toward negative viewpoints of the testing program. The views about the positive aspects of NAPLAN varied across socio-demographic groups, but there were more similar views concerning the negative views of NAPLAN. As predicted by the social-cultural and institutional performance models, those living in metropolitan areas and those from high SES groups tended to possess stronger positive views. Political party preference was a strong predictor of the perceptions about NAPLAN. Overall, this study offers possible explanations for the underlying mechanisms explaining sub-group differences in attitudes toward large-scale standardized national testing.”

The findings of the present study may provide educators and policymakers with a better understanding of how the NSW public may have perceived NAPLAN and why certain sub-groups may hold more favorable or more critical views on the national testing program. More varied sub-group differences were noted in perceptions about the positive aspects of NAPLAN while fairly consistent and widespread negative views were shared among different socio-demographic subgroups. Discontent that is frequently expressed by members of the general public can erode the program’s sustainability (Jacobsen, Snyder, & Saultz, Citation2014; Wichowsky & Moynihan, Citation2008). To gain stronger support and trust from members of the general public, educational authorities and policymakers may need to more openly and clearly inform them about both the benefits and negative consequences of national testing.

Lee, J., Lee, J. S., & Ellis, N. (2023). Public Opinion About National Large-Scale Student Assessment: A Case of NAPLAN. Educational Assessment28(3), 137–155. https://doi.org/10.1080/10627197.2023.2215977

____________________________________________________________________

“In the midst of the debate surrounding the question of whether Australia’s National Assessment Program: Literacy and Numeracy (NAPLAN) test is high-stakes, it is evident that children’s own accounts of their experiences remain sparse. This paper describes the findings of a case study which documented the experiences of 105 children across two Catholic primary schools in Queensland serving different socio-economic status (SES) communities. Analysis of the data revealed that these teachers and principals did not experience NAPLAN as high-stakes. However, the data suggested that the children experienced the tests within a confusing context of contradictions and dissonances emanating from multiple sources; receiving little, if any, clear and consistent information regarding the purpose and significance of NAPLAN. While the children’s responses were varied, many reported NAPLAN as a negative experience, with some constructing the test as high-stakes. These constructions ranged from personal judgement or sense of letting their families down, to failure, and less commonly, grade retention and school exclusion. Some Year 3 children had also constructed good results as vital to future prosperity. These constructions bring into question the assumption that because NAPLAN is designed to be a low-stakes test, that children will necessarily experience it in this way.”

Howell, A. (2017). ‘Because then you could never ever get a job!’: children’s constructions of NAPLAN as high-stakes. Journal of Education Policy32(5), 564–587. https://doi.org/10.1080/02680939.2017.1305451

_____________________________________________________________________

“Standardised testing regimes, including the National Assessment Program Literacy and Numeracy (NAPLAN) in Australia, have impacted on relationships between and within schools, and on teachers’ work and on pedagogies. Previous analyses of the effects of NAPLAN have been generated outside of the test situation: frequently through attitudinal surveys and qualitative interviews. This article takes as its point of departure two intensely affective events associated with the NAPLAN test day itself. These events erupted in two qualitative studies of students’ schooling experiences: a study of students’ experiences of NAPLAN and a study of students’ experiences of student voice at school. We ask, after Deleuze and Guattari, What can a NAPLAN test do? Exploring the entangled corporeal (physical and embodied) and incorporeal (psychic and subjectivating) wounds effected in and through these events, we analyse the dynamic constitution and re-constitutions of ‘at risk’ categorisations. While the NAPLAN test is not claimed to cause physical and psychical injury, we argue that standardised test conditions, in these singular events, are inextricably entwined with the formation of particular students’ schooled subjectivities.”

Mayes, E., & Howell, A. (2017). The (hidden) injuries of NAPLAN: two standardised test events and the making of ‘at risk’ student subjects. International Journal of Inclusive Education22(10), 1108–1123. https://doi.org/10.1080/13603116.2017.1415383

___________________________________________________________________

“This paper engages, Sara Ahmed’s theorising on ‘the uses of use’ to frame an analysis of the hidden, embedded effects of standardised testing policy that have become normative practice/s in Queensland, Australia. It (re)examines data from an ethno-case study into the datafication of assessment and learning over one school year, in primary and secondary schooling contexts, to understand the uses of the National Assessment Program – Literacy and Numeracy (NAPLAN) in a new, critical light. We explore schools’ contemporary uses of NAPLAN – intended or otherwise – to demonstrate how the policy effects of NAPLAN have become insidiously submerged within the daily practices in schools. Drawing on interviews with 27 teachers and seven school leaders, classroom and staff meeting observations, and artefact data, we reveal the invisible yet profoundly altering presence of NAPLAN and its consequences. Specifically, we analyse the ways in which NAPLAN practices, structures and technologies are both hidden and yet manifestly altering as a) practices that disappear into their uses, becoming unidentifiable and routine; and b) practices that follow well-used pathways that further embed particular uses. We counter rhetoric of NAPLAN normativity and complacency, instead demonstrating that its current uses, while not originally intended, are insidious and profound.”

As Ahmed (2019) suggests, a way that an object is used is shaped by the requirements of its use, and in the case of NAPLAN, improved performance as defined by narrow measures such as NAPLAN scores, greater accountability and data visibility requirements have shaped the way/s that NAPLAN is deployed by educators and policymakers. As an influential policy driver in schools, and a mechanism used to inform staff decisions and resource allocation, we have shown that NAPLAN rationalities now exist as a normative, embedded element of school-level decision-making. As NAPLAN envelopes teachers’ practice, it becomes both normative and unremarkable, and so its reach becomes harder to distinguish between what is teaching work, and what is NAPLAN-oriented teaching work.

Rafaan Daliri-Ngametua, Stephanie Wescott, and Amanda McKay (2024), On the uses and use of NAPLAN: the hidden effects of test-based data-centric. accountabilities. JOURNAL OF EDUCATION POLICY, 39(6), 843–860.

https://doi.org/10.1080/02680939.2023.2273499

________________________________________________________________

Module-Bottom-Button-A rev

Module-Bottom-Button-B rev

Module-Bottom-Button-C rev2

candid-seal-gold-2024.png