The Illusion of Knowing by David Didau @LearningSpy | http://www.learningspy.co.uk/ The author of ‘What if everything you knew about education was wrong’ BROCHURE TITLE About David Didau In 2011, frustrated by the state of education David Didau began to blog. His blog, The Learning Spy, is ranked as the second most influential education blog in the UK and has won a number of awards. In addition to this Didau has also authored a number of books including ‘The Perfect English Lesson’, ‘The Secret of Literacy’ and his most recent book ‘What if everything you knew about education was wrong?’ which explores the idea that much of what happens in schools is based on unexamined assumptions. As well as working as a freelance writer, speaker and trainer, Didau is currently employed by Swindon Academy as an internal consultant advising on curriculum design, teaching and literacy. 2 | The Illusion of Knowing Contents Contents 3 Summary 5 Introduction 4 Ignorance isn’t bliss; it’s scary 4 If we collect enough data, can we predict the future? 4 Can you have too much data? 6 Perverse incentives 8 What should we be collecting and analysing? 8 How can we guard against bad data? 9 O brave new world! 9 3 Summary “The cost of bad data is the illusion of knowledge” Stephen Hawking Working groups are looking at reducing the administrative burden on teachers. At the same time we are told that data on learners is at the heart of helping them to achieve their full potential. This means more data, doesn’t it? And more data means even more administration? Not according to David Didau who argues that we don’t need more data, we need better data, and that with technology this does not necessarily mean more workload. Introduction The more data schools collect, the better they will understand their students, right? Well maybe not. Much of the data schools collect distorts how we think and warps decisions we make about what to teach. The illusion of knowing is comforting, but maybe we’d be better off confronting the uncomfortable truth that we know far less than we suppose. As we find stuff out, we reveal new boundaries to what we know. As the island of knowledge grows, so does its shoreline and beyond that shore is a vast ocean of ignorance. Nate Silver said “Even if the amount of knowledge in the world is increasing, the gap between what we know and what we think we know may be widening.” Ignorance isn’t bliss; it’s scary The vast, black ocean of our ignorance makes us uncomfortable. Most of the time, we occupy the centre ground of the island of knowledge in which things are safe – only rarely do we venture on to the shoreline to peer into the void. Much easier to deal with, and concentrate on, what is already known. And when we’re certain of what we know we feel even more secure. Certainty makes us feel safe; no matter how bad it is, we can cope. But there are some problems with certainty. On the whole, we’d prefer people were wrong than unsure. We punish politicians or school leaders for acknowledging that they don’t know and are much happier when they’re decisive, confident… and mistaken. We prefer erroneous information to no information at all. Entertaining uncertainty is uncomfortable; we fight to feel sure. The thing about data is that it provides us with a powerfully seductive sense of certainty. Data 4 | The Illusion of Knowing occupies the high ground on our island of knowledge; it sorts and quantifies what we know into handy, bite-size chunks which tell us what to believe and how to act. We make decisions based on the data we collect in the belief that, because it’s readily available, it’s more likely to be accurate. Of course, sometimes the information we’re able to draw to mind might be accurate, but sometimes it won’t be. Validity and Reliability We could inoculate ourselves with a better understanding of some key statistical concepts. For the data we produce to be useful it must be both valid and reliable. Briefly, reliability relates to the accuracy and precision of what’s being measured, while validity refers to the depth and scope of what’s being measured. Without these considerations we cannot be sure the data supports us in making the inferences we want to make. The reliability of data is just as important as its validity. We need to trust that the data we gather allows us to make meaningful inferences about what’s going on in schools. Too often, high stakes decisions are made on insufficiently reliable evidence. Essentially, the higher the stakes, the more important precise measurement becomes. And if precise measurement isn’t possible – as is all too often the case – we should avoid rushing to judgement. If we collect enough data, can we predict the future? There are clear limits to the power of data, but there seems to be a common misconception that if only we could collect a large enough sample size we would be able to predict anything. We could use data to make predictions by imagining that our prediction is true and then working out the probability of getting the results we’ve got. Sadly though, we end up imagining our data is precise and accurate, and then trying to work out the probability that our prediction is true. a.I have made a prediction. What is likelihood that my prediction is true given the results I’ve observed? b.I have observed some results. What is the likelihood that my prediction is true? This confusion is at the heart of everything that goes wrong. Let’s reframe the problem with more familiar terms: a.I ate some muesli this morning. What’s the probability I had breakfast? b.I ate breakfast this morning. What’s the probability I had muesli? Suddenly, it’s much easier to see where we’re likely to go wrong. Added to this is the fact that the future is a moving target. We might be able to predict the likelihood of a set of results if nothing else changed, but that’s not how reality works. No amount of data can accurately describe the complexity of human systems. Maybe the data we collect at Christmas might help us to predict GCSE results with a reasonable amount of accuracy, but data collected 5 years previously is much less reliable as the margin of error only grows with time: inspection frameworks change; external assessment are changed mid-year; teachers leave, new teachers arrive; students experience spurts and declines. Reality gets in the way of our careful calculations. Because we’re so poor at dealing with uncertainty, we struggle to accurately forecast the future. All prediction is based on probability; these probabilities are risks which we calculate by analysing historical data. Uncertainty though, is different to risk. A risk describes the possible range of all known outcomes and tries to calculate the likelihood of each occurring. All too often, we don’t even know that an outcome is possible, let alone the probability of it occurring. Data management systems can only tell us what has happened in the past. If the future is different to the past – as it inevitably will be – Before we start collecting data, we might do well to check its validity and reliability. Ask yourself these 4 questions: 1. How much of what I want to measure is actually being measured? 2. How much of what I did not intend to measure is actually being measured? 3. What are the intended and unintended consequences of making these measurements? 4. What evidence do I have to support my answers to the first three questions? 5 Schools are awash with data. We need to know how to interpret data clearly, appropriately and fairly. any forecast we make will be wrong. We have school dashboards, the National Pupil Database, Raise Online, School Performance Tables, the Fischer Family Trust, all churning out floods of figures which must be funnelled appropriately. Many, if not most, schools these days employ a full time Data Manager to nourish and nurture all these numbers. As with so much else in life, the having of a thing is not its purpose. Analysing spreadsheets and graphs can be like gazing, dumbly, into a crystal ball. We need to know how to interpret what the data tells us. And, perhaps more importantly, we need to know what it can’t tell us. If data is to be used to make decisions about setting or grouping pupils according to ability, diagnosing what individual pupils know, promotion or retention, or measuring the effectiveness of instruction. For instance, numbers assigned to graded lesson observations are entirely and utterly subjective, but they provide the comfortiing illusion of hard data. As Robert Coe, Professor of Education at Durham University says, “If we were to use the best classroom observation ratings, for example, to identify teachers as ‘above’ or ‘below’ average and compare this to their impact on student learning we would get it right about 60% of the time, compared with the 50% we would get by just tossing a coin.” Many school leaders have been seduced into the 6 | The Illusion of Knowing easy certainties of grading lesson observations, aggregating the grades, and then proudly declaring that teaching in their school is 80% good or better. But all this means is that they like 80% of their lessons. There is no objective standard to compare this data against. Assigning numerical values to our preferences and biases gives them the power of data, but they’re still just made up. Now that Ofsted have accepted that attempting to grade lessons is invalid and unreliable, many school leaders are anxious about how to go about holding teachers to account. This is an understandable dilemma but one with no easy answers. What we must remember is that while its inescapably human to make judgements, we should avoid high-stakes decisions based on bad data. Can you have too much data? Data comes with other costs. When the government issued its call for evidence on teachers’ workload, there was an unprecedented response. Thousands of teachers poured out their angst at the uneccessary and unproductive uses to which their precious time was being put and 56% of respondents cited the recording, inputting, monitoring and analysis of data as the chief culprit. We might mourn the passing of national curriculum levels, but they are an example of made up data at its worst To do this we need new systems which work for teachers rather than creating work for them to do. Frustratingly, much of the data we so dutifully collect tells us as little as lesson observation grades. Teacher and statistician Jack Marwood puts it starkly: “Just about all of the internal progress-tracking ‘data’ and external test ‘data’ which is used in English education could actually be more correctly described as Cargo Cult data – that is to say, it has the appearance of real, measurable data, without fulfilling any of the requirements of being statistically valid, approximatelymeasured, error-accepting, countable data from which one could reasonably draw inferences. The fundamental problem with education data is that it cannot be subjected to close statistical scrutiny.” Contrary to what most people believe, teacher assessments cannot reliably track pupil progress and external tests can’t be used to measure educational achievement. Even the most accurate assessment is guesswork, and the constraints of designing and marking tests mean that our best guesses are often wrong. Tests provide a snapshot or sample of a pupil’s actual knowledge, skills and understanding of a given subject. On a different day, a pupil is highly likely to get a different mark and quite likely to be awarded a different grade. Dylan Wiliam warns that up to 32% of students have been given the wrong National Curriculum levels and urges us to be aware that “the results of even the best tests can be wildly inaccurate for individual pupils, and that high-stakes decisions should never be based on the results of individual tests.” Someone (often a teacher) is asked to assign a numerical value to students’ work on a regular basis. We then pore over these tea leaves as if they are an objective reality instead of someone’s best guess about how a student may have performed in a particular task on a particular day. And then we write reports saying with absolute certainty that ‘Emily is a 4b in writing’ and ‘Isaac is a 5c in maths’. But what does this actually mean? What can we know about what Emily or Isaac can actually do? These numbers provide the illusion of knowledge. And on this illusory foundation we build a house of cards with which to hold schools and teachers to account. We stumble into the same sort of traps when setting targets for students. We tell them they need to know their target grades as if they are cast iron certainties. But while they may not be simply plucked from the air like lesson observation grades, they’re based on statistical probabilities that may have some validity when applied to large cohorts but which are reduced to meaningless nonsense when applied to individuals. Currently, many companies and organisations are desperately seeking ways to replace levels with… more levels. It doesn’t have to be this way – the removal of statutory assessment levels allows schools to focus on what children can actually do. 7 What should we be collecting and analysing? Valuable insights can be gained by having a system which is fit for purpose and provides a bespoke lens. Perverse incentives The problem for schools is that data management systems end up wagging the teaching dog. Mass collection is a wasteful fetish, but in the brave new world of freedom from government shackles on assessment, schools are free to track and monitor what students can actually do rather than quantify vague, generic skills. When we try to work out whether students fit the level descriptors for level 4 in writing we might miss the fact that they don’t understand subject-verb agreement; rather than collecting bogus data on the percentage of students achieving level 5 in mathematics we could track whether students are able to accurately compare negative numbers. Often, our response to a perceived difficulty is to offer incentives. Students misbehaving and not working hard enough? Vivo Miles! Teachers not working their fingers down to the bloody bone? Performance Related Pay! If no incentive is offered we’re in the terrifying position of simply relying on altruism. But, if people are properly incentivised, the reasoning goes, they will act with motivation, determination and efficiency. We tend, unerringly, to respond to incentives by doing what is in our best interests. We respond to the letter of the initiatives rather than their spirit; we ignore the intention and focus solely on the incentive. Because schools and teachers are judged on the results of the children in their charge, a lot of time is spent testing and recording how well students can answer questions in external written tests on which they will be examined, 8 | The Illusion of Knowing The middle of the school year – March 2nd – should be central to every teacher’s understanding of the relative abilities of the children in a class. This is data worth using. but this isn’t an outcome anyone really wants or values, but it is a logical response to a perverse incentive. Economist, Rolf Dobrelli offers this advice: “Good incentive systems comprise both intent and reward.” Clearly this is sage advice, but how so to do? When managing change, we tend to focus on problems. We look at what is preventing people from behaving in the way we want them to behave rather than focusing on those instances of success. The best way to change behaviour is to focus on these positives instead of all the frustrating negatives. An accountability system that allows people to either input numbers they’ve made up or wrestle data into meaning something it was never intended to mean is doomed to fail. But what’s worrying is that many schools, teachers, parents and children aren’t even aware of the failure. What should we be collecting and analysing? One of the most important but widely neglected data sets we should be collecting and analysing is the age of students within a year group. This is a useful, robust measurement which actually affects what a child is able to do. Too often, high ability groups are simply the older children in a O brave new world! class, with lower ability groups made up of the younger children. But when it comes to learning and progress we must be mindful that these are too complex to be reduced to numbers. We can certainly record indicators of children’s progress against what teachers think they should have learned based on the content that has been covered. Clearly this is highly subjective, as long as no-one tries to abuse this information by feeding into high-stakes accountability systems, then it might be useful. It’s also possible to track the acquisition of the fundamental knowledge vital for pupils to make progress in education. For example, we can monitor pupils’ knowledge of number bonds and times tables, since this a clear and unambiguous test of recall of factual information. We should also track pupils’ knowledge of the alphabet, phonemes and graphemes and punctuation marks for those who have yet to master writing. Attempting to track what children can do is difficult, but monitoring progress is possible, and teachers can assess whether pupils are using certain key concepts correctly. What we can’t do is assign numbers to make sense of what’s going on. O brave new world! With the freedoms schools now have, it’s increasingly possible to resist the accountability club and concentrate on what you actually want to know about the pupils in your school. Valuable insights can be gained by having a system which is fit for purpose and provides a bespoke lens. The market-leading behemoths with the constant ‘add-ons’ and work arounds, just How can we guard against bad data? Next time someone shows you a spreadsheet, try asking these questions: •If the data is the solution, what’s the problem? •Is there a different way of interpreting the data? •How can I verify the quality of the data I’m being shown, and what are the margins of error? increase workload. The prize is out there for anyone flexible enough to provide data managements systems that can integrate with what schools really want and what they’re already doing. Maybe this is an argument for providing opportunities for schools and teachers to design their own apps which can be brought together under a powerful operating system? There are some amazing homemade systems out there for collecting and analysing truly useful data. Colin Hegarty’s award winning website, Hegarty Maths, sequences all the skills students need to be successful in mathematics, provides instruction, assessment and accurate tracking of what students can do, and intervenes to either plug gaps or provide further challenges. At Michaela School in Brent, teachers ensure students’ mastery of key curriculum content by setting challenging multiple choice questions and then giving students weekly quizzes in each subject. Their bespoke data management system tracks the time students spend answering questions, identifies common misconceptions, logs their progress and challenges them to do even better. Dr Chris Wheadon’s No More Marking system is an exciting development in considering how we can generate data which liberates rather than constrains. Although still in the pilot phases of development, the system asks teachers to upload essays which are compared and placed into a rank order. The system doesn’t rely on computer programmes or complex algorithms, instead teams of subject experts (PhD students working for the sheer love of it) read a couple of •What are the limitations of these data – what don’t they show? •How is the data likely to affect my decision making? What would I do differently if I didn’t have the data? •How was the data acquired, and is it valid? By being more critical of the data with which we’re presented maybe, just maybe, we can avoid some of the potential pitfalls associated with the over abundance of data in schools 9 Section headline over one or two lines essays and decide which one they like best. Each essay is judged by a number of different experts and their subjective opinions are aggregated. There are no vague or over-complicated markschemes to interpret and teachers can select any scale – 1-20, 1-100 they wish; the system will record the aggregate of the experts’ marks accordingly. This then allows teachers and students to have meaningful discussions about why an essay has scored a particular mark to drive precise, generalisable feedback on how performance might be improved. The purpose behind the new breed of management information systems is to put schools in control of their data, not the other way around. For instance, Advanced Learning’s Progresso system helps schools put “the right 10 | The Illusion of Knowing information in the hands of the right people at the right time”, how many schools can say the same of the system they use? These examples underline that we should only collect data we can use to give us insights and to flag up problems we might otherwise miss, but they also reveal the expertise and effort required to produce a management system fit for purpose. What schools really need are data management tools to meet them where they are and provide them with the wherewithal to maintain and improve standards of pupil achievement. References >> David Thomas, “Be scared of the myth of big data” http://davidthomasblog.com/2015/09/ be-scared-of-the-myth-of-big-data/ >> Dylan Wiliam, Reliability, validity, and all that jazz http://eprints.ioe.ac.uk/1156/2/ Wiliam2001Reliability3long.pdf (p 3) >> Government response to the Workload Challenge, February 2015 https://www.gov. uk/ government/uploads/system/uploads/ attachment_data/file/415874/Government_ Response_to_ the_Workload_Challenge.pdf >> hegartymaths.net >> Nate Silver, The Signal and the Noise >> nomoremarking.com >> Robert Coe, Cesare Aloisi, Steve Higgins and Lee Elliot Major, What makes great teaching? Review of the underpinning research p. 3 >> Rolf Dobrelli, The Art of Thinking Clearly p. 57Marwood, Jack, Data by Numbers in Didau, D, >> Vivo Miles is a commercial rewards system intended to work a little like Air Miles. Pupils gain rewards for good behaviour and you know what points make? That’s right. Prizes. >> Marwood, Jack, Data by Numbers in Didau, D, What if everything you knew about education was wrong? (2015) p. 337-338 11 About Advanced Advanced provides management solutions, Facility and Progresso to schools around the world. By basing its solutions on the latest cloud technology and appreciating the everchanging nature of education, it works with its customers to deliver systems which are primarily focused on learners and improving outcomes for them. To get an insight into this approach and to learn more about the work Advanced is doing with some of the UK’s leading schools, please contact us on 0330 060 2199 or learning. [email protected] More information w oneadvanced.com t +44(0) 8451 605 555 e [email protected] Ditton Park, Riding Court Road, Datchet, SL3 9LL Advanced Computer Software Group Limited is a company registered in England and Wales under company number 05965280, whose registered office is Ditton Park, Riding Court Road, Datchet, SL3 9LL. A full list of its trading subsidiaries is available at www.oneadvanced.com/legal-privacy. BROCHURE TITLE
© Copyright 2026 Paperzz