X Close

IOE Blog

Home

Expert opinion from IOE, UCL's Faculty of Education and Society

Menu

Imagine there's no GCSEs… It's easy if you try

By Blog Editor, IOE Digital, on 25 October 2013

John White
This month’s OECD report has linked the poor literacy and numeracy skills of our 16-24 olds, compared with those in other countries, with a lower level of social mobility. One factor in holding back social mobility has been the way in which the more affluent have used school examinations to entrench their dominance in higher education and in professional jobs.
A way of countering this dominance would be to remove its instruments: school exams themselves. These may have made sense when they first became popular around the 1850s, but do they in 2013?
They made sense then for the rising middle classes who wanted their sons to have interesting careers. Before that, patronage had been the norm. Exams were held to provide a fairer and more objective alternative.
Are they worth retaining now? Social unfairnesses apart, we know what an obstacle they are to a worthwhile education for students in their last years at school. We know about the anxiety they cause, the overwork, the narrowing of the curriculum, the teaching to the test, the training in question-spotting, examiner-bluffing and other morally dubious habits.
What do schools get out of examinations? Universities and employers are the beneficiaries. Schools are their handmaidens, relieving them of work and expense. Apart from kudos for some in the league tables, schools get mainly grief.
Imagine a world without GCSEs and A levels. How much broader and richer education could become! The school could come back to its proper task of creating in every school leaver a passion for learning – rather than hacked-off attitudes among its exam failures and instrumental attitudes to learning among its successes.
Would anything be lost in an exam-free régime? What about selecting candidates for higher or further education and employment?  We need to probe imaginative alternatives. Colleges and employers should take the lead on devising their own filtering devices, as long as these do not disadvantage less privileged applicants. But schools can also play a part.
The 1970s and 1980s witnessed the growth of records of achievement, or student profiles. These enabled schools to provide ongoing accounts of progress in different areas. They were not tied to the framework of discrete subjects within which examinations tend to be conducted. They recorded progress on non-academic fronts, including practical and out-of-class activities, within the school and outside.
In those days profiling existed alongside conventional examinations. It could now replace them. It could give every school leaver a record of his or her all-round attainments, presentable both to other educational establishments and to employers. Our new digital age has opened wider horizons, with children and parents as well as teachers contributing to web-based records, not only with text but also with still and moving images. This already happens in enlightened schools.
I know all this is anathema to those who will tell us what a subjective, corruption-prone system profiling is. Examinations are so much fairer, so much more accurate in their assessments. This is why colleges and employers trust them – and why they have for over a century provided a ladder for young people from all social classes to reach the top.
We should not accept this. Examining something is submitting it to thorough investigation. Do the minutes an examiner allots, say, to a paper on Europe since 1815 really add up to an examination? Digital cumulative records can give a much fuller picture; and of the student not only as a lone learner, but also as a co-worker – and not only in this area or that, but also as a whole person. And who is to say that there are no ways of coping with teacher bias – or, indeed, other problems around profiling?
The ladder argument raises another point. The more affluent have long accepted examinations as their route to university and professional jobs. They have been happy, at the same time, with some poorer but bright youngsters joining them. The ladder is said to provide equality of opportunity.
But the story says nothing about equipping everyone for a decent life. The ladder for the lucky ones, too, has always been propped up against a smart building with escalators inside to take the affluent up to the upper storeys. Michael Gove’s policy on examinations has upgraded the escalators and made the ladder more rickety.
Examinations are an excellent device for keeping a hierarchical social order more or less intact. If we are more interested in introducing every young person to the delights and rewards of learning, we have to look elsewhere.

Assessment: levelling off?

By Blog Editor, IOE Digital, on 17 June 2013

Chris Husbands

On Thursday, OFSTED reported that a quarter of those who secured Level 5 at the age of 11 did not get top grades at GCSE; I blogged on it here.  On Friday, government solved the problem: it abolished National Curriculum levels.  If only all educational challenges could be solved so easily.  I picked up the news on a train returning from a day working with headteachers in Somerset and we were powering across the Somerset Levels, which felt somehow appropriate.

The eight national curriculum levels in use now derive from Paul Black’s Task Group on Assessment and Testing (TGAT) Report of 1988.  It is a sophisticated report, still worth reading, which tried to devise a national assessment system precise enough to give information about pupil progress but flexible enough to recognise that pupil learning is not simply cumulative.  As Black observed, “no system has yet been constructed that meets all the criteria of progression, moderation, formative and criterion-referenced assessment” (para 5).  TGAT’s 10-level scheme, drawing together test results, teacher assessment and teacher moderation was intended to secure “ information about … progress … in such a way that the individual child’s progress can be related to national performance” (para  126).

The implementation of TGAT was – as is always the case – partial.  The recommendation that “wherever schools use national assessment results in reports for evaluative purposes, they should report the distribution of pupil achievements” (para 84) was lost as progression was mapped through a level-based system .  Assessment at 14 was never implemented as intended.  Working groups each generated statements of attainment at each level in each subject – Science alone had seventeen attainment targets, so 170 statements of attainment.   TGAT’s proposals were, as reform proposals always are, sharply criticised.  Some who wanted “simple” tests disliked the attempt to draw external and teacher assessment into a single 10-level structure.  Others thought it would lead teachers to think of pupils as numbers.  But level-based assessment was rapidly adopted, not least because in areas of the school system it was working with the grain of thinking.  Substantial work had been done on graded assessment in languages, GCSE operated levels of response mark schemes, and primary mathematics schemes often had a level-based structure. Lord Dearing’s review of the national curriculum in 1994 moved from statements of attainment to more generic level descriptors, and lopped levels 9 and 10 off TGAT’s structure.

Since 1994, the use of levels has become embedded.   Some teachers – wrongly – assign levels to every piece of work, which misses the point about generic descriptors.  Better practice happens where  teachers have used levels to inform assessment and planning and sharpen the relationship between, say, level 5 writing and level 5 speaking in English,  or progression in History between levels 5 and 6.  Precisely because levels are not disaggregated, they have informed thinking about what performance looks like across different cognitive domains and how domains relate to each other.  Primary and secondary classrooms often feature posters using  pupil-friendly language to help pupils identify how to consolidate learning and move on.  The application of level descriptors itself involves professional judgement.  Black’s aspirations for the way the TGAT structure could shape progression and formative assessment have been substantially achieved. 

Ironically, it is what TGAT envisaged as criterion-reference assessments related to national performance that have caused more difficulty.  Such has been the focus on key hinge points of the structure – the importance attached to the proportion of pupils securing a Level 4 at Key Stage 2, for example – that schools have engaged in frenetic activity to push pupils over the ‘next’ threshold for key assessments, so that a higher level is achieved, but not always securely.  ‘Sub-levels’ –disaggregated progression within levels – have generated data of dubious reliability.  Partly for this reason, many secondary schools re-assess pupils soon after they arrive, and, perhaps inevitably, decide to place greater emphasis on these tests than on national curriculum levels.

The government’s announcement that levels are being abandoned suggested that “outstanding schools and teaching schools” would lead the way in devising alternative ways of tracking progress. More likely will be an expansion of external tests, like those from NFER, CEM or from commercial test and publishing companies – News Corporation has been buying into the market in the USA. National benchmarking will still exist through end of Key Stage tests; without levels, these will presumably generate a numerical score, though TGAT warned (para 10) that where external tests are detached from school assessment practice they tend to be “narrow in scope and employ a limited set of techniques….[and] rarely assist normal teaching work”. 

Profusion of local and commercial assessment puts at risk the real long-term achievement of  TGAT:  its contribution to raising standards across the school system.  It made expectations are clearer – from Carlisle to Canterbury, from Newquay to Newcastle.  Some won’t mourn levels, especially those who have seen them misused.  But the level structure focused attention on what pupils need to learn, learning outcomes, and on progression towards those outcomes.  

TGAT recognised that “many schools have established common systems for use throughout the school”, but teachers’ use of assessment was “limited by the range of  instruments available, by the lack of time and resources required to develop new instruments, and by the lack of external reference and a basis for comparison”, so the “needs of the less able, or the competence of the most able, have hitherto been under-estimated” (paras 7-8).  And that, more or less, brings us back to OFSTED’s concerns.

 

Bold GCSE reform depends on clarity of purpose

By Blog Editor, IOE Digital, on 17 September 2012

Chris Husbands
After a summer of controversy in examinations, the government appears to be on the verge of a decisive policy announcement on the future of GCSEs. If the Mail on Sunday is to be believed,  the announcement will mark a significant shift in assessment practice for pupils at the age of sixteen: a greater focus on terminal assessment, more stretch and challenge at the higher grades, and greater discrimination at the upper end of the attainment distribution – even if, as Adam Creen points out in his sharply written commentary, most of the aspects of the announcement trailed on Sunday are already features of current practice. The report was clear on some issues: in the new proposed GCSEs, the article declared, “as few as one in ten will get the top mark”, although faith in the reporter’s numeracy skills was somewhat undermined by the claim later in the article that “as few as five per cent may get Grade 1”. Too bad for the reporter that re-sits are to be strictly limited.
There is a widespread consensus that GCSEs need serious reform. They were introduced in 1986, when they effectively completed the task of the Raising of the School Leaving Age (RoSLA) in 1973, by requiring all students to complete year 11 in school.  But their function as a “school leaving” examination in a system in which the vast majority of learners will be in education and training up to the age of 18 appears unclear. They are inflexible: as Kevin Stannard from the Girls Day School Trust comments,  “they are too chunky and require a certain number of recommended hours, so schools can only fit a limited number into the curriculum”. Widespread media concerns about “grade inflation” have dominated parts of the media,  though as Jo-Anne Baird, professor of assessment at Oxford told the Education Select Committee,  these concerns are more difficult to pin down to hard evidence. The extent to which they have driven not only the assessment of pupils but also the accountability of schools means that any number of perverse incentives have appeared in the system: the focus of effort on the C/D borderline means that insufficient attention is paid to the long tail of poorly performing pupils,  so that we have a longer tail of low performance than many other countries. A bold reform, building a consensus around the aims and purposes of upper secondary assessment and deriving the form and nature of assessment from these purposes could command widespread support.
Bold reform depends fundamentally on clarifying the main purpose of assessment.  It’s possible to design an assessment system principally to identify and rank top performing pupils. Kenya has a system somewhat like this and every year the Kenyan press hunt out the top performing girl and boy. It’s possible to design an assessment system to identify those apparently most suited to particular types of subsequent learning, whether academic or vocational. Hermann Hesse’s (in my recollection, almost unreadable)  novel The Glass Bead Game explored one possibility, whilst Michael Young’s much misunderstood account of the Rise of the Meritocracy thought hard about the long-term consequences of such a system.  It’s possible to design a system to assess the performance of schools – though the gaming of thresholds and entry rules which this produces provides ample evidence of Campbell’s law  that “The more any quantitative social indicator is used for social decision-making,… the more apt it will be to distort and corrupt the social processes it is intended to monitor.”  It’s possible to design an assessment system to recognise and reward the achievement of all who reach a certain level, irrespective of how many actually do. What it’s not possible to do is to design a system which does all these things at once without introducing  confusion and “noise”.
There are some big lessons about assessment reform from experience around the world:  it is complicated.  It has lots of unforeseen consequences. It takes time. Done properly,  as it has been in high performing jurisdictions as different at  Finland and Hong Kong, it can drive higher standards for all, shaping professional expectations and engendering commitment across the system. But it depends on clarity of purpose in building a framework up which all learners can climb.

A new binary divide will not solve the real challenges

By Blog Editor, IOE Digital, on 26 June 2012

 Chris Husbands
The blogosphere is bristling with responses to the Daily Mail’s story about the possible return of O-levels. I began my teaching career in the early 1980s. One of my abiding memories – and bitter frustrations – is that each year, 15-year-olds who had been cajoled, exhorted and motivated to keep going through CSE courses simply left at Easter,  and never  turned up for their exams.  They saw no real point in turning up to complete an examination which they thought of as dead end – with no progression route and little labour market validity. In this respect, at least, they were showing themselves as pretty shrewd labour market economists.
So I was part of a generation of teachers which welcomed the introduction of GCSE in 1986. The Conservative secretary of state for education,  Sir Keith Joseph, was determined that the new examination would be “tougher, because it would demand more of pupils; would be fairer because pupils would be judged by what they could do and not how they compared to someone else; and would be clearer because everyone would know what had been tested.” Sir Keith’s aim was to get 80-90% of pupils up to the level previously thought to be average. As  Caroline Gipps,  at the time a senior member of staff at the Institute of Education, pointed out,  on norm referenced tests such as O-level, there is no point in trying to get every pupil to achieve an above average score, since, by definition, such tests are designed to have half the population scoring above and half below the mean.
By this measure, GCSE has been an enormous success. Performance rose:  41% of pupils scored A-C grades in 1988, but by 2011 the figure was 69%. School staying on rates increased sharply:  they had been 36% in 1979, but rose to 44% in 1988, 73% by 2001 and almost 80% by the end of the decade.  GCSE completed the 1973 work of RoSLA – the Raising of the School Leaving Age from 15 to 16. By and large, GCSE achieved the levering up of performance which Joseph had expected.  
But none of this makes it unproblematic.   One of the challenges was explained as long ago as 1994 by Caroline Gipps. GCSE used criterion-referenced assessment, and so “as the requirements become more abstract and demanding, so the task of defining the performance clearly becomes more complex and unreliable”. Put differently, it becomes more difficult to design assessment criteria which work at both extremes of the performance range. But it is not impossible, and assessment experience here and elsewhere suggests it can be done by ensuring a common core of curriculum entitlement, and a sufficiently varied and stimulating curriculum diet that there are opportunities for all young people at all levels to experience success. 
A second challenge was not foreseen in 1988, and followed the annual publication of examination results focusing on 5 A/A*-C performance from the early 1990s: although  in technical terms a GCSE pass was a grade G or better, league tables reinforced the idea – imported from an old O-level equivalence – that the cusp performance was at Grade C. There were thus incentives for schools to focus their effort on moving marginal performance at grade D up to grade C, and it became no easier to motivate a pupil on track for a grade G to improve by one grade than it had been to motivate the CSE students of the early 1980s.
The difficulty for the nation is that neither of these problems will be solved by introducing a new binary divide into qualifications,  even if,  as the leaked reports of DfE thinking suggest,  the revamped O-level is “targeted” at the top 75% of the attainment range rather than the 60% target group for O-levels and CSE. There are two reasons. The first is that any system which designs in a selective process at the beginning of examination courses has a backwash effect: a divided system at 14 means making selection decisions by 13. The analysis by the Financial Times’s Chris Cook  suggests that this will have a sharply differential effect in different parts of the country. Moreover, with any threshold there will be errors about mis-classifying pupils into the “wrong” route, closing down opportunities and dampening motivation. Ben Levin and Michael Fullan, writing about education system reform, warn that “literacy and numeracy goals must include higher-order skills and connections to other parts of the curriculum, such as science and the arts, to avoid the curriculum… becoming too narrow and disengaging”.  
The second reason is that the most serious performance challenge we face as a nation is to do what our major competitors are doing and to seek to bring all young people up to Level 2 performance by the time they leave compulsory education. Given that the education participation age is rising to 17 and then to 18, the challenge is a curriculum rather than an assessment one:  how do we secure high-quality, labour-market valid outcomes for all young people? That’s a question of curriculum design, educational quality and learner motivation.