Tag Archives: learning difficulties

The Optima Reading Programme by Dr Jonathan Solity: Does it Provide Optimal Results? A Paper by Dr Marlynne Grant

Recently Dr Jonathan Solity reported on his critical analysis of the Year 1 Phonics Screening Check.1 This has thrown a spotlight on his own programme for teaching reading, which was developed first as ERR (Early Reading Research), later known as KRM, and currently as Optima.

Having watched the video on the Optima website (2016) (2) http://optimapsychology.com/find-out-more/introductory-video/ I was struck by the many similarities to the systematic synthetic phonics (SSP) teaching with which I am familiar:

  • Whole class teaching
  • Results showing no gender differences or vulnerable group differences
  • Key skills of synthesising and segmentation
  • All achieving success from the start i.e. from Reception
  • Successful with low achievers and high achievers who are stretched
  • Pacy
  • Works on writing as well as reading
  • Use of modelling
  • Little and often teaching in daily doses
  • Very similar teaching each day with a little bit extra added (cumulative)
  • Establishes love of books and literature
  • Complete engagement
  • Increases confidence and enjoyment
  • Parents are impressed as they see structure and progress

Optima video from cris on Vimeo.

On the video, Solity stresses that Optima is underpinned by psychological theory. But Optima does not have the monopoly on incorporating psychological learning theory into its teaching:

For example, one of the government approved SSP programmes* (3) lists the following:

  • Whole class teaching
  • Reinforcement and repetition are built in
  • Use of active recall – students are encouraged to make active attempts at recall
  • Oral work
  • Interactive teaching which engages the children rather than them having to work individually on work sheets
  • Lively teaching. The lesson has a good pace, which helps to manage behaviour and focus students’ attention
  • Multisensory teaching. It integrates what you see (letters) with what you hear (sounds) and what you do (articulating sounds and words, clicking fingers, phoneme fingers, robot arms, manipulating sound and syllable cards, writing sounds and words from dictation)
  • Frequent rehearsal on the ‘little and often’ principle
  • Develops fluency and mastery in learning which is vital and will ensure later success with reading comprehension and writing composition
  • Direct instruction (modelling) is used: I do, we do together, you do.

*Sound Discovery Manual, p4 (3)

The Optima video claims that its programme is ‘different from everything else around’ – so what are these differences? It appears to differ from government approved SSP programmes with regard to the following:

  • Lack of decodables
  • Teaching High Frequency Words (HFWs) as sight words
  • Teaching fewer grapheme-phoneme correspondences (GPCs)

Decodable Reading Books

It is not clear whether Solity has conducted research to justify his decision not to use the sort of decodable reading books often used now with SSP programmes. In these books a high percentage of words (average of 81% in the SSP study reported below) embody only the GPCs so far taught, with divergences in the few remaining words being explicitly taught.

One longitudinal study of SSP (4) compared outcomes with and without decodable reading books (http://bit.ly/2coAoKP ). The study showed a gain of an extra 5 months of reading age for a cohort of 85 Reception children when decodable reading books were introduced for the first time, and when other teaching variables remained the same. Subsequently in following years 433 Reception children used decodable reading books as their first experience of reading books for themselves alongside a SSP programme, using the phonics they had been taught. Over time these pupils were able to demonstrate an impressive start and sustained attainments, even those from vulnerable groups.

To give further information on this study: Grant reported on whole cohorts of Reception children being taught synthetic phonics over an eight-year period. At the beginning of the study, during one year, the whole cohort of Reception children was not given decodable reading books, but continued to use the Oxford Reading Tree books employed by the school at that time, which were based largely on look-and-say and whole language principles. The 90 Reception children achieved an average of 12 months ahead of chronological age for reading and 17 months ahead for spelling at the end of Reception. The following year decodable reading books were written and used for the first time. 85 Reception children averaged 17 months above chronological age for reading, a gain of an extra 5 months of reading age compared with the previous year. Averaged over 6 years with SSP and decodable books, 433 children were 16 months ahead for reading and 17 months ahead for spelling. (4, p18) Cohorts of these Reception children were tracked through to their KS1 and KS2 English SATS where comprehension and writing as well as decoding skills were assessed.

There is other supportive evidence for the use of decodable reading books, for example, the work of Juel and Roper/Schneider, 1985 (5) as referenced in: http://www.iferi.org/iferi_forum/viewtopic.php?f=2&t=469 (6). In addition there are the summaries of research into decodable text discussed in: http://www.iferi.org/iferi_forum/viewtopic.php?f=2&t=554 (7).

Teaching HFWs as Sight Words

Optima proposes teaching the 100 most frequently occurring words in English as whole words by sight, whether or not these words can be read using the GPCs already taught. “These were taught as whole words with no reference to any phonemes within the words …..” (8) (Shapiro and Solity, 2008). The 100 words were based on word lists and “determined as the optimal number by Solity, McNab and Vousden, 2007 (9); Solity and Vousden, 2009 (10); Vousden, 2007 (11).

The practice of learning HFWs by sight is in direct contrast with the view of the psychologist and researcher Professor Diane McGuinness (12) and with the government’s advice (13):

Research has shown, however that even when words are recognised apparently by sight, this recognition is most efficient when it is underpinned by grapheme-phoneme knowledge”. “What counts as ‘decodable’ depends on the grapheme-phoneme correspondences that have been taught up to any given point”. “Even the core of high frequency words which are not transparently decodable using known GPCs usually contain at least one GPC that is familiar. Rather than approach these words as though they were unique entities, it is advisable to start from what is known and register the ‘tricky bit’ in the word.” (13, pp 15-16).

Diane McGuinness (12) has pointed out that introducing multiple strategies (such as learning HFWs by sight) at an early stage of reading instruction will be “mutually contradictory and will confuse rather than assist young readers”. (http://www.rrf.org.uk/archive.php?n_ID=112&n_issueNumber=51)

The dual system of reading instruction proposed by Optima would not meet the DfE revised core criteria which defined the key features of an effective systematic synthetic phonics teaching programme (13). The Department strongly encouraged heads and teachers to consider the revised core criteria when making decisions about the quality of commercial programmes and the suitability of them for their particular schools and settings. The first of these criteria was ‘phonics first and fast’ as a programme’s prime approach to decoding print.

Optima follows an explicit dual route to reading instruction: teaching phonics along with learning HFWs by sight. Although SSP programmes teach phonics ‘first and fast’ they can also teach ‘tricky’ words but again through phonics as recommended in Letters and Sounds (L&S) (14). In my own programme, common ‘tricky’ words are introduced in a natural drip-fed way through sentence work, right from the very beginning of instruction. (3 and 17)

Shapiro and Solity (15) compared the effectiveness of L&S (“which teaches multiple letter-sound mappings”) implemented in Reception with ERR (“teaches only the most consistent mappings plus frequent words by sight”), and then followed up outcomes at the end of the second and third years of schooling. They found the two programmes equally effective in broad terms with ERR being more effective with children with poor phonological awareness.

In her blog on sight words (http://readoxford.org/guest-blog-are-sight-words-unjustly-slighted) (16), Professor Anne Castles reviewed the Shapiro and Solity 2015 study (15) and agreed with them that teaching frequent words by sight did not appear to interfere with phonics learning in the ERR programme.

Shapiro and Solity (15) made some positive comparisons of ERR compared with L&S. However, would these comparisons hold up if ERR were compared with other SSP programmes? Shapiro and Solity suggested that in L&S some children may not have fully grasped the concept of segmentation and blending in the absence of print before moving on to segmentation and blending of letters. They pointed out, in contrast, that ERR continued to teach segmentation and blending in the absence of print in every whole class lesson. Is this true for any government ‘approved’ SSP programme? In at least one other SSP programme (3 and 17) practice with oral segmentation and blending precedes segmentation and blending of letters in every lesson.

Shapiro and Solity (15) also maintained that ERR provided more practice with ‘tricky’ common words and that it might be more beneficial for L&S children simply to learn these by sight and gain regular practice with them, instead of attempting to analyse their sound structure. However, other SSP programmes may provide more explicit teaching and practice for HFWs. For example, in one other SSP programme there are resources (18) provided to support the analysis of sound structure in HFWs. Specially prepared common words, sentences and texts make it easier for children to practise sounding and blending them, so that the children begin to be able to read the words without overt sounding and blending. Thus, they start to experience what it feels like to read some words automatically. In addition, blending and segmenting of common ‘tricky’ words takes place in every lesson alongside more easily decodable words, providing the ongoing regular practice that Shapiro and Solity recommended.

At present there appears to be insufficient evidence to support the Shapiro and Solity view that only ERR has potential benefits for children at risk of developing reading difficulties. They suggest that an optimal programme should explicitly teach children to use two strategies: sight recognition and phonic decoding. However, to date, they have not published data which compares ERR with SSP programmes other than L&S, particularly where those programmes use reading books with a high level of decodability and which explicitly support teaching for HFWs.

The Grant longitudinal study (4) used a ‘phonics first and fast’ approach but also taught HFWs through sentence work and through a phonics route. Impressive results were achieved even with vulnerable groups of pupils who were at risk of developing reading difficulties and with higher achievers who were stretched (0% below Level 3B, 6% Level 3B, 94% Level 4+, 65% Level 5 at KS2 English SATs).

Number of GPCs Taught

I have some sympathy with the Solity view regarding the number of grapheme-phoneme correspondences (GPCs) which should be taught. How extensive should the advanced code be? Solity questions, “whether it is worth teachers spending a great amount of time making sure pupils learn all 85, rather than concentrating on the most frequent ones and then building pupils’ vocabulary.” (1)

Solity is following the ‘principle of optimal information from rational analysis’ and aims to teach the “optimal number of GPCs(8). However these do seem to be quite limited. Children were taught 61 high frequency mappings between graphemes and phonemes (26 letters; 5 vowels modified by ‘e’; 30 letter combinations) and “where multiple mappings exist between phonemes and graphemes – only one GPC was taught”. (8)

Teaching in this way would require more words having to be taught explicitly should they contain a GPC not covered in the 61 high frequency ERR mappings. Would these words also be taught by sight?  Such words would be in addition to the ‘tricky’ common words taught as “unique entities” by sight. In SSP programmes a greater number of words would be decodable as more GPCs would be taught and words with unusual GPCs would be blended from what is known and from registering the ‘tricky’ bit’ in the word as recommended by DfES. (14, pp15-16).

Learning a large number of words by sight in this way could place a strain on memory to which there is a limit. Diane McGuinness reported the average visual-memory limit for whole word units as approximately 2,000 (19). A good English dictionary contains from 250,000 to 500,000 words, which sounds like a huge challenge for those individuals needing to memorise whole words.  Whereas those able to use a more comprehensive alphabetic code would be at an advantage and more able to work out pronunciations using their pre-existing phonics.

In her blog (16), Professor Castles suggested that some pupils in the Shapiro and Solity study (15) were possibly confused by being exposed to the multiple alternative sound mappings (GPCs) in L&S rather than to sight words in ERR.

What about the strain on memory of learning GPCs? Is there a memory limit to the number of GPCs that can be taught explicitly? According to the literature this limit is high. Victor Mair, Professor of Chinese Language and Literature, claims there is a natural upper limit of approximately 2,000-2,500 to the number of sound-symbol units (in our case GPCs) which most individuals can tolerate (20). Also note the reference in http://www.rrf.org.uk/archive.php?n_ID=173&n_issueNumber=59).  I think we are safe in saying that this limit far exceeds the demands of all SSP programmes. Even those SSP programmes which teach a large and very comprehensive alphabetic code for English are unlikely to be teaching more than 100-200 GPCs. Some SSP programmes will be teaching fewer and a number will be concentrating on even fewer. In Sound Discovery (3) every effort is made to teach the advanced code in the simplest and most straightforward way in order to decrease confusion and minimise overload.

From the perspective of limits to memory, one cannot assume that learning 100 sight words and a reduced set of GPCs as in ERR is less strain on memory than learning more GPCs and constantly rehearsing and practising common ‘tricky’ words through phonics in a SSP programme.

There may be a value in teaching a more comprehensive alphabetic code than Optima in a systematic way as recommended by the Year 1 Phonics Screening check (21), even if, in practice, a greater emphasis is placed on the most frequently occurring GPCs. We simply do not know whether programmes teaching less code are less effective or more effective when there is no comparative research.

Reducing Difficulties

In the Optima video (2), Solity reported that his programme had reduced reading difficulties from 20-25% to 2-3%.

It is perhaps not surprising that such positive results and views have been reported, given the systematic way the Optima programme appears to have been introduced and delivered in schools, the commitment of all staff, including senior management, and the ongoing support from the Optima team. However we do not know whether even better results could have been achieved if they had incorporated some of the features discussed above and which are found in good synthetic phonics teaching (viz. decodables, not teaching HFWs by sight, teaching a more comprehensive alphabetic code – even if then concentrating on the most frequently occurring).

Dr Solity quoted a percentage of remaining ‘difficulties’ as 2-3% with Optima. In the Sound Discovery study (4) we were able to achieve just over 1% of moderate reading difficulties in 2004. Only one child out of the 3 form entry at Year 6 achieved less than Level 4 English. 94% of pupils achieved Level 4+ and 65% achieved Level 5. It is not clear how Solity defines ‘difficulties’ as they relate to his quoted 2-3%. In the Grant study the only pupil who did not achieve Level 4 gained a Level 3B English which is not a severe literacy difficulty. This pupil had complex and severe learning difficulties and he was followed up into his secondary school. He was reported as, “holding his own in mainstream classes in Year 7; he had made good gains in reading and spelling and could understand complex vocabulary in the curriculum. He was able to be de-statemented in Year 9” (4, p19).

Real Books

The Optima video stresses the importance of ‘real’ books in increasing the vocabulary and language comprehension of pupils.  The Optima team did not use the sort of decodable reading books matched to their order of teaching GPCs often used now by SSP programmes.

In contrast SSP schools comply with the Simple View of Reading (22) which identifies two distinct processes in learning to read: ‘word recognition’ and ‘language comprehension’. Many SSP programmes have a strong language comprehension strand using structured, decodable reading materials which aim not only to give practice with decoding but also to develop vocabulary and comprehension.

In addition, in SSP schools, ‘real’ books and rich literature are used alongside decodable books for adults to read to children and to share with them. The aim is to establish a love of books and literature and to increase confidence and enjoyment. Children taught in this way pick up reading quickly. They become enthusiastic and confident about their reading. They are more able and willing to engage in the world of reading around them and take advantage of incidental phonics practice in the environment. They are also more able and willing to access a wide range of texts and literature themselves.

Conclusion

The Optima video was impressive and I am not surprised that the schools were achieving good results with a programme and teaching which the majority of us could endorse. However, in my view, similar outcomes could be achieved and even surpassed in schools which are committed to following rigorously a good quality systematic synthetic phonics programme, which uses books which are decodable at a high level (in this instance at the 81% level as mentioned on page 2) and teaches HFWs with attention to what is, and is not, decodable in them.

 

Dr Marlynne Grant

Registered Educational Psychologist

Author of the government ‘approved’ systematic synthetic phonics programme Sound Discovery

Committee member of RRF

October 2016

 

References

 

  1. Richardson, H. (2016). National phonics check ‘too basic’. BBC News, 16th September 2016. Education and Family, available online at: http://www.bbc.co.uk?news.education – 37372542 ; and British Education Research Association (BERA) press release, available online at: https://www.bera.ac.uk/bera-in-the-news/press-release-children-can-pass-phonics-test-without-extensive-phonic-knowledge

 

  1. Optima Video (2016). Optima Psychology, available online at http://optimapsychology.com/fi nd-out-more/introductory-video /.

 

  1. Grant, M. (2000). Sound Discovery Manual. Synthetic Phonics Ltd., www.syntheticphonics.net .

 

  1. Grant, M. (2014). Longitudinal Study from Reception to Year 2 (2010-2013) and Summary of an earlier Longitudinal Study from Reception to Year 6 (1997-2004). The Effects of a Systematic Synthetic Phonics Programme on Reading, Writing and Spelling. Paper presented to ResearchEd Conference, London, 2014. Available online at http://bit.ly/2coAoKP .

 

  1. Juel, C. & Roper/Schneider, D. (1985). Reading Research Quarterly, 18. Also in Adams, M.J. Beginning to Read: Thinking and Learning about print, Bradford Books, pp275-280.

 

  1. International Foundation for Effective Reading Instruction (IFERI), 2015. Is there a role for predictable texts in reading instruction? Available online at: http://www.iferi.org/iferi_forum/viewtopic.php?f=2&t=469 .

 

  1. International Foundation for Effective Reading Instruction (IFERI), 2015. The multi-cueing reading strategies and ‘Is reading about getting meaning from print’? Available online at: http://www.iferi.org/iferi_forum/viewtopic.php?f=2&t=554 .

 

  1. Shapiro, L. & Solity, J. (2008). Delivering Phonological and Phonics Training within Whole Class Teaching. British Journal of Educational Psychology, 78, part 4, pp 597-620.

 

  1. Solity, J., McNab E. & Vousden, J. (2007). Is there an optimal level of sight vocabulary to teach beginning readers? Unpublished data.

 

  1. Solity, J. & Vousden, J. (2009). Reading schemes vs. real books: A new perspective from instructional psychology. Educational Psychology, Volume 29, Issue 4.

 

  1. Vousden, J. (2007). Units of English spelling-to-sound mapping: a rational approach to reading instruction. Applied Cognitive Psychology, Volume 22, Issue 2.

 

  1. McGuinness, D. (2004). A response to teaching phonics in the National Literacy Strategy. RRF (Reading Reform Foundation) Newsletter 51. Available online at: http://www.rrf.org.uk/archive.php?n_ID=112&n_issueNumber=51 .

 

  1. Department for Education (DfE) (2010). Phonics teaching materials: core criteria and the self-assessment process. Crown copyright. Available online at: https://www.gov.uk/government/uploads/system/uploads/attachment_data/file/298420/phonics_core_criteria_and_the_self-assessment_process.pdf .

 

  1. Department for Education and Skills (DfES) (2007). Letters and Sounds: Principles and practice of high quality phonics. Notes of guidance for practitioners and teachers. DfES Publications, pp15-16. Available online at: https://www.gov.uk/government/publications .

 

  1. Shapiro, L. & Solity, J. (2015) Differing effects of two synthetic phonics programmes on early reading development. British Journal of Educational Psychology, Volume 86, Issue 2, pp182-203.

 

  1. Castles, A. (2016). Guest Blog: Are sight words unjustly slighted? Read Oxford, University of Oxford. Available online at: http://readoxford.org/guest-blog-are-sight-words-unjustly-slighted .

 

  1. Grant, M. (2007-2009). Sound Discovery Big Books of Snappy Lesson Plans at Step 1, Step 2, Step 3A, Step 3B and Steps 4-7. Synthetic Phonics Ltd. www.syntheticphonics.net

 

  1. Grant, M. (2014). Sound Discovery High Frequency Words, Version 2. Synthetic Phonics Ltd. www.syntheticphonics.net .

 

  1. McGuinness, D. (2004). Growing a Reader from Birth, W.W. Norton.

 

  1. Daniels, P. and Bright, W. (1996) (Editors). The World’s Writing Systems. Oxford & New York: Oxford University Press, p 200, and referenced in:

http://www.rrf.org.uk/archive.php?n_ID=173&n_issueNumber=59 .

 

  1. Department for Education (2011). Year 1 Phonics Screening Check.

https://www.gov.uk/government/policies/reforming-qualifications-and-the-curriculum-to-better-prepare-pupils-for-life-after-school/supporting-pages/statutory-phonics-screening-check .

 

  1. National Primary Framework for Literacy and Mathematics (DfES) (2006). Crown Copyright.

A New Paper by Professors James W. Chapman and William E. Tunmer on Reading Recovery

IFERI is delighted to be able to share with you a brand new paper by Professors James W. Chapman and William E. Tunmer, from the Institute of Education at Massey University, New Zealand.

This paper was presented, by invitation, at the 39th Annual Conference of the International Academy for Research in Learning Disabilities (IARLD), Vancouver, Canada, July 8, 2015. Professor James Chapman has been a Fellow of IARLD since 1983.

IARLD (International Academy for Research in Learning Disabilities) is an international professional organization dedicated to conducting and sharing research about individuals who have learning disabilities. Fellows of IARLD include premier scientists, educators and clinicians in the field of learning disabilities throughout the world.

For convenience, some extracts and conclusions from the paper are published as part of this blog post. To open or download the complete paper, simply click the title below.

The Literacy Performance of ex-Reading Recovery Students Between Two and Four Years Following Participation on the Program: Is this Intervention Effective for Students with Early Reading Difficulties?

 James W. Chapman and William E. Tunmer

Sustainability of Gains Made in Reading Recovery

Considered together, the PIRLS results for 9-year-old children who had received RR in Year 2, the enrolment data for students receiving support from RT:Lits, and the two New Zealand studies on the sustainability of RR outcomes for discontinued children, show that RR simply has not achieved its primary goals in New Zealand. Clay’s avowal that RR would “clear out of the remedial education system all children who do not learn to read” (Clay, 1987, p. 169), and the RR New Zealand’s website claim that RR operates as an “effective prevention strategy against later literacy difficulties” and, therefore, “may be characterised as an insurance against low literacy levels” (www.readingrecovery.ac.nz/reading_recovery), are without foundation.

Why Does Reading Recovery Fail to Result in Sustainable Gains?

We have argued elsewhere (Chapman et al., 2015) that the effectiveness of RR interacts with where children are located on the developmental progression from pre-reader to skilled reader. Because of limited knowledge of print at the outset of learning to read, and/or developmental delay in acquiring the phonological awareness skills that are essential for learning to read successfully (e.g., Pressley, 2006; Snow & Juel, 2005; Tunmer, Greaney & Prochnow, 2015), a large proportion of young struggling readers operate at low developmental phases of word learning, which Ehri (2005) described as pre-alphabetic and partial-alphabetic phases. Delayed readers who are still in these phases, typically those students who struggle the most with learning to read, will not be able to grasp the alphabetic principle and discover spelling-to-sound relationships on their own or in a program that emphasizes text rather than word level instructional approaches. These students will require more intensive and systematic instruction in phonemic awareness and phonemically based decoding skills than what is provided in typical RR lessons.

What Should be Done to Improve the Effectiveness of Reading Recovery?

There are serious shortcomings and much-needed improvements in several aspects of RR, including the theoretical underpinnings of the program, the assessment battery which fails to include measures of phonological processing skills, the specific instructional strategies emphasized in the program (e.g., the multiple cues approach to word identification), the manner of program delivery (one-to-one versus instruction in pairs), and the congruence between classroom literacy instruction and the RR program.

Regarding the issue of congruence between classroom literacy instruction and RR, the program was originally developed to complement regular whole language classroom literacy instruction in New Zealand. Clay (1993), nevertheless, claimed that RR was compatible with all types of classroom literacy programs, but she offered no evidence in support of this claim. To test this belief, Center et al. (2001) investigated whether the efficacy of RR varied as a function of the regular classroom literacy program. They compared the effects of RR in “meaning oriented” (i.e., whole language) classrooms and “code-oriented” classrooms (i.e., those that included explicit and systematic instruction in phonological awareness and alphabetic coding skills). Their results indicated that at the end of the second year of schooling, children in the code-oriented classrooms (regular and RR students combined) significantly outperformed children in the meaning-oriented classrooms on measures of phonological recoding, reading connected text, and invented spelling, as well as on a standardized measure of reading comprehension. Overall, however, Center et al. (2001) reported that the RR students in both types of classrooms failed to reach the average level of their peers on any of the literacy measures. These findings clearly contradict Clay’s (1993) claim that the regular classroom context does not differentially affect the literacy performance of RR children.

Although regular classroom literacy instruction influences the effectiveness of RR, the most serious shortcoming of the program is the differential benefit at the individual level. The program may be useful in the short term for some struggling readers but not others, especially those struggling readers who need help the most. More intensive and systematic instruction in phonemic awareness and phonemically-based decoding skills is likely to be required than what is normally provided in RR lessons for those who struggle most with learning to read, and for any gains made in RR to have a lasting effect (Iversen, Tunmer & Chapman, 2005; Tunmer & Greaney, 2010).

Slavin et al. (2011) found reading programs for younger children that had less emphasis on phonics, including RR, showed smaller effect sizes than those programs that included phonics. They noted that RR is the most extensively researched and used reading intervention program in the world, but that the outcomes were less than might be expected. Further, Slavin et al. observed that the overall effect size for 18 studies involving paraprofessional or volunteer tutors using structured and intensive programs was about the same as the effect size for RR studies (+0.24 vs. +0.23), despite the very intensive training that RR teachers receive.

Conclusion

The RR program remains largely un-revised in its instructional approach despite clear evidence showing that claims about RR being an insurance against on-going literacy difficulties are without foundation. The New Zealand Reading Recovery website continues to assert the effectiveness of RR; assertions that are not supported by the New Zealand Ministry of Education’s own data (national monitoring reports and PIRLS), or by the two independent studies undertaken in New Zealand on students two to four years following successful completion of the program. If the RR program is not changed to reflect contemporary scientific research on reading interventions, it should be dropped and replaced by a more contemporary, research-based, reading intervention approach, together with more effective literacy instruction in children’s first year of schooling.

See also:

Excellence and equity in literacy education: the case of New Zealand. W.E. Tunmer & J.W. Chapman (eds.) (June, 2015). Basingstoke, UK: Palgrave Macmillan.

http://www.palgrave.com/page/detail/excellence-and-equity-in-literacy-education-william-e-tunmer/?K=9781137415561

Special exam arrangements for dyslexia veering out of control

Special exam arrangements for dyslexia veering out of control

Julian Elliott, Durham University

The new English Literature GCSE might be contravening the 2010 Equality Act, according to concerns raised by a teacher. This raises serious questions about which disabilities should lead to students being given special dispensation in exams.

The new GCSE, which will be taught from September 2016, requires students to remember a number of poems and analyse at least one. Mary Meredith, a teacher and blogger who works with pupils with special learning needs, argued in an open letter to education secretary Nicky Morgan that those with dyslexia may be disadvantaged unfairly because they typically experience problems with verbal memory. She has asked for the exam to be adjusted for dyslexic pupils.

There are a number of problems with this argument. The usage of the term “dyslexia” has expanded to the point that it has now lost much of its explanatory value. Initially it was used to describe very rare cases in which people could make little or no sense of the written word. Now it has become the diagnosis of choice to describe individuals exhibiting one or more of a wide range of cognitive difficulties involving areas such as memory, speed of processing, attention, concentration, analysis and synthesis, organisation and self-regulation – controlling oneself and one’s actions.

While not all those with reading difficulties experience memory problems, and not all those with memory problems struggle with literacy difficulties, there is clear evidence that a greater proportion of pupils with reading difficulties encounter problems with short-term or working memory. However, working memory appears not to be a particularly powerful predictor of reading difficulties. Short-term memory essentially involves holding information in your mind for short periods of time, for example trying not to forget a phone number while you struggle to find a pencil. If this information slips away, it is typically gone forever – what is called “catastrophic loss”.

Working memory is very similar but also involves the process of doing something with the information while we are holding onto it, for example, undertaking a complex calculation in our heads. These types of memory are rather different from those involved in remembering information that has been stored for a longer period – such as the ability to recite a poem from memory or the reasons behind the Russia revolution. The evidence that poor readers have particular difficulty with this latter form of “long-term” memory is much weaker, with inconsistent findings from research studies.

Qualifying for special assistance

We cannot assume that someone with a diagnosis of dyslexia has one particular type of memory problem or, indeed, any memory problem at all. There is also the question about how we should respond to those people who have poor memories but are not considered to be dyslexic. It would be a travesty to automatically diagnose all those with working memory problems as dyslexic even though, anecdotally, this seems to be an increasing tendency in university disability services and in some school contexts.

Students’ memory abilities will be normally distributed within the general population – this is typically the case for any cognitive process. So how can we best respond to those who argue that their memory difficulties disadvantage them in examinations? One solution could be to establish a national screening programme to identify all those who encounter various memory difficulties and who might be deemed to need special arrangements. Clearly, this is not feasible or desirable.

Alternatively, if we conclude that all examinations discriminate against those with memory problems, there would be a logical argument to be made for scrapping examinations that place any burden upon recall. Open book exams, where a student has the relevant information in front of them, is one possible solution, but it might also be challenged on the grounds that those with working memory difficulties are still disadvantaged. As I hope it is becoming clear, there is an inherent weakness of logic here.

The go-to diagnosis.
Dyslexia via Groenning/www.shutterstock.com

In reality, examinations are not just designed to test a student’s memory. They tap the ability to utilise knowledge and understanding in ways that are underpinned by a whole range of intellectual abilities. These include processes such as memory, speed of processing, attention, concentration, analysis and synthesis, organisation and self-regulation. But herein lies a major problem: that’s a very similar list (see the second paragraph above) that is often used to describe those with a dyslexic disability and who seek and are offered special assistance or modifications in exams.

Special arrangements were derived to give everyone a fair opportunity to access and engage in examinations. Where someone has a physical disability it makes obvious sense that they should be able to record their responses in alternative ways, for example, by dictation or a modified keyboard. Where a student is blind, it is wholly appropriate that a person or a keyboard should read the questions (and the candidate’s responses) back to them. In such situations, the individual is getting no additional assistance that gives them an advantage over other candidates.

However, where the person’s difficulty involves the same processes as those being used to differentiate between candidates’ academic performance, for example, remembering detail, or being able to marshal and express a complex argument in a time-constrained period, we run the risk of helping some while disadvantaging others.

Where the underlying problems concern cognitive processes such as memory, processing speed or attention, the preferred strategy should not be to modify examination conditions. Instead, it should be to assist students to develop strategies that can assist them to perform as well as possible, targeting additional resources towards the provision of workshops to improve relevant study skills.

Out of control

Currently, understandings and practices around this issue are so confused that the existing systems are getting increasingly out of control. Universities, for example, are provided with assessments and recommendations from privately funded educational psychologists that are often difficult to challenge.

The fact that many of these reports are severely flawed – for example, many still employ the now widely discredited IQ discrepancy criterion for dyslexia – seems not to be considered problematic by politicians and civil servants, whose focus appears to rest primarily around the increasing cost of provides resources for more and more diagnosed dyslexic students.

If individualised resourcing and exam accommodations are to be provided for particular cognitive weaknesses, we need a more sophisticated understanding of exactly which difficulties, at what level of severity, are appropriate for special assistance. What should be more widely understood, however, is that these processes are significant in the way we currently differentiate between students’ academic performance. So to assist one person, but not another, in this fashion throws up serious questions about equity and fairness.

The Conversation

Julian Elliott is Professor of Education and Principal of Collingwood College at Durham University.

This article was originally published on The Conversation.
Read the original article.