Evidence for Learning has published its evaluation report on MiniLit, a small group, phonics-based, program for struggling Year 1 readers. Early support for struggling readers is crucial to putting them on a path to reading achievement, which is a key predictor for future academic success. With an ongoing public debate about effective reading approaches, this report is an important contribution to the discussion.
The trial found that MiniLit did not have an additional impact on passage reading, but that there was evidence of significant improvement in foundational skills, particularly Letter Sound Knowledge which sustained even six months after the program was completed. It also suggested greater gains for students who attended 80% or more of the sessions. Overall the evaluation has a low security rating due to concerns about the test measure used and the level of change the trial was set up to detect, meaning it needs to be treated with caution.
MiniLit is delivered in school outside of regular class by teachers or paraprofessionals trained as MiniLit tutors, to small groups of up to four students. In this trial it was tested with Year 1 students in the bottom 25% of reading in nine NSW public schools. Half the students were assigned to 80 unique one-hour lessons over 20 weeks. The other half of students received the school’s usual support for struggling readers. All students’ reading levels were tested after the MiniLit program concluded and the results were compared. This randomised controlled trial (RCT) was conducted by the Centre for Community Child Health and the Centre for Program Evaluation and commissioned by Evidence for Learning as part of its Learning Impact Fund.
Evidence for Learning Director Matthew Deeble said: “This trial highlights the complexity of building reading skills and the challenge of measuring its development in the critical early years of reading. The primary measure of reading, selected at the outset of this trial, was too ambitious for these students meaning the findings need to be treated with caution. But in the process, valuable knowledge has been generated about MiniLit’s positive impact on the development of skills that lead to confident reading and the more likely benefit when receiving the full program.
“It shows the value of independently conducted trials, that publicly report on the differences in achievement between trial and comparison groups, evaluate the steps along the way and calculate the costs to get there. The evaluation report, its methods and results are available for everyone to learn from. But most importantly, accompanying evaluation resources have been developed for educators.
“Creating this kind of knowledge is not easy. It requires expert evaluators with a range of skills to produce high quality research working in partnership with an education system and its schools. But none of it can happen without the commitment of the program developer who opens their program up to a new level of scrutiny. I applaud MultiLit for their outstanding contribution to advancing our collective understanding about reading development. And I thank Murdoch Children’s Research Institute, Melbourne Graduate School of Education and the NSW Department of Education.”
Through its Learning Impact Fund evaluations, Evidence for Learning is increasing rigorous evidence on the cost and effectiveness of educational programs and practices in Australian schools. MiniLit is the third randomised controlled trial, following two other trials testing numeracy programs.
The evaluation report and accompany resources are publicly available on the Evidence for Learning website, https://evidenceforlearning.org.au/lif/our-projects/minilit/.