Skip to main content

Hattie’s Effect Size: A pseudoscience or critics just being critics?


Hattie’s meta-mata analysis that culminated in the publication of his most influential work of Visible Learning (2009), and later updated to include more studies, has been hailed as the “holy grail” for educators and education leaders around the world. In particular, his effect size of instructional practice interventions has had the lion’s share of his work. Hattie considered that if schools set the effect size at 0 then “virtually everything works, and so we need to shift the question from “ what works in education” to “what works best in education”. Hattie’s meta-meta analysis of more than 800  meta-analyses studies comprising 50,000 studies (later included more 1500 meta-analyses) revealed that the baseline of the effect size that schools should start from is not 0 but 0.4, termed as the “hinge point”. In other words, for medium to large effect sizes on student achievement, the effect size of an instructional practice should be o.4 and above. This does not mean that we need to discard all positive effect sizes below o.4. It means that what works best is o.4 and above. Hattie also published a list of 150 influences on student achievements based on their effect sizes, now being collective teacher efficacy as the number one influencer on students achievement, teacher feedback, student and teacher expectations, student-teacher relationships, and  feedback for teachers among medium to large effect sizes.
It is worth noting  also that when considering an effect size, other factors need to be considered as well, like resources, effort, and time put into an instructional strategy. For example, if a school opted from smaller classes, which has a medium effect size, they need to consider the resources, money, and time they need to accommodate to have such a a medium effect size, which by comparison to teacher feedback (done well) the latter makes much more sense as it yields higher effect size with much less effort and resources.
However, Hattie’s effect size validity has been under numerous attacks lately, accusing it of being skewed. I was aware of such criticism before, but it has become more salient lately through some publications and discussions on Twitter.






This Twitter conversations were due to Dr. Robert Salvin’s (currently Director of the Center for Research and Reform in Education at Johns Hopkins University and Chairman of the Success for All Foundation) post John Hattie is Wrong  where he accuses Hattie of bias study findings by accepting “ the results of the underlying meta-analyses without question” as “most meta-analyses accept all sorts of individual studies of widely varying standards of quality”. He adds that
To create information that is fair and meaningful, meta-analysts cannot include studies of unknown and mostly low quality. Instead, they need to apply consistent standards of quality for each study, to look carefully at each one and judge its freedom from bias and major methodological flaws, as well as its relevance to practice. A meta-analysis cannot be any better than the studies that go into it. Hattie’s claims are deeply misleading because they are based on meta-analyses that themselves accepted studies of all levels of quality.

Towards the end of this post, Dr. Slavin refers to his work on Evidence Syntheses (www.bestevidence.org) as reliable and unbiased. Was he trying to promote his work? Does he feel Hattie’s work and commercialization of his work have overshadowed his work on reliable meta-analysis?

Hoever, Dr. Salvin’s post was not the only criticism, earlier in 2017 an academic article by Jerome Bergeron, published in McGill Journal of Education, and titled  HOW TO ENGAGE IN PSEUDOSCIENCE WITH REAL DATA: A CRITICISM OF JOHN HATTIE’S ARGUMENTS IN VISIBLE LEARNING FROM THE PERSPECTIVE OF A STATISTICIAN  likened Hattie’s research to “a fragile house house of cards that quickly falls apart”. He claims that Cohen’s d (Hattie’s measure of effect size) simply cannot be used as a universal measure of impact. He even accused Hattie, his works, and his followers and believers as blind and it falls into the promotion of pseudoscience (reminds me of NLP).
To believe Hattie is to have a blind spot in one’s critical thinking when assessing scientific rigor. To promote his work is to unfortunately fall into the promotion of pseudoscience. Finally, to persist in defending Hattie after becoming aware of the serious critique of his methodology constitutes willful blindness.

He says that Hattie, though not afraid of numbers, has created many errors, but the most two salient ones are:
  1. Miscalculation in Meta-analyses
  2. Inappropriate baseline comparisons
He says that Hattie does not use statistically sophisticated formulas, that he summarizes by calculating averages and standard deviations, the latter he did not use, that he uses bar graphs not histograms and uses a formula that converts a correlation to Cohen’s d without being heedful of the validity of his conversion.

He insists that effect size, despite the absence of a unit, is a relative measure that provides a comparison to a set, group, or baseline population, even if it may be implicit. He says that to compare two independent groups is not the same as comparing grades before and after an intervention implemented with the same group. Hattie’s comparisons are arbitrary and he is completely unaware of it.
He also points to an error of arbitrariness of Hattie’s barometer. He says that in addition to mixing multiple and incompatible dimensions, Hattie confounds two distinct populations: 1) factors that influence academic success and 2) studies conducted on these factors.
He concludes his article with a call for researchers in education and education departments to consult keen statisticians instead of doing it themselves.

Another fierce criticism on Hattie’s work (and Hattie himself) was termed the “cult of Hattie” in another academic article by Scott Eacott in 2017, titled School leadership and the cult of the guru: the neo-Taylorism of Hattie. In this article, Eacott contends that Hatti’s word in Visible Learning “has become not only the latest fad or fashion, almost to the point of saturation, but reached a level where it can now be labelled the ‘Cult of Hattie’. In which claims that Hattie has deliberately discarded context, a very powerful influence on student achievement. Eacott’s attack, unlike Bergeron’s, does not attack the statistical errors in Hattie’s work but claims that Hattie did three things to build is Visible Learning empire (He is localizing Hattie’s influence in Australia):
1- Hattie uses specific temporal conditions that enabled the rise of Hattie’s work across the nation (Australia)
2- Unlike past attempts at pedagogical reform, Hattie’s work has provided school leaders with data that appeal to their administrative pursuits>
3- Hattie articulated a new image of school leadership (him being their savior)

What do you think? Is Hattie’s “Holy Grail” the wrong one? a pseudoscience? or are these critics, among many others, are just being well, Critics?

Comments

Popular posts from this blog

198 METHODS of Nonviolent Actions to Instill Change

In 1972, Gene Sharp produced  a brochure listing 198 methods on nonviolent actions to instill change. Since then, his methods have become a blueprint for nonviolent actions around the world. I publishing it here in the hope that this might spark done ideas for change. Practitioners of nonviolent struggle have an entire arsenal of “nonviolent weapons” at their disposal. Listed below are 198 of them, classified into three broad categories: nonviolent protest and persuasion, noncooperation (social, economic, and political), and nonviolent intervention. A description and historical examples of each can be found in volume two of The Politics of Nonviolent Action, by Gene Sharp. THE METHODS OF NONVIOLENT PROTEST AND PERSUASION Formal Statements                     1. Public Speeches                 ...

Capturing Solutions for Organizational Learning and Scaling Up

The World Bank has published a much needed guidebook for organizations on how to document operational experiences for organizational learning and knowledge sharing . It also discusses the significance of organizational capabilities at two levels: Enabling Environment for Knowledge Sharing and Technical Skills. The  publication asks a simple, yet important, question: Is your organization missing important lessons from its operational experiences? This step-by-step guide shows you how to systematically capture such knowledge and use it to inform decision making, support professional learning, and scale up successes. The captured lessons--knowledge assets, the central element needed for learning--are consistently formatted documents that use operational experience to answer a specific question or challenge. The guide describes how to create and use knowledge assets in five steps: (1) identify important lessons learned by participants, (2) capture those lessons with text or multimedi...

In Lebanon, a Socio-culturally Fueled Cancer Is on a Steady and Alarming Rise

Last week, we had a visitor from a nearby village in Bekaa, Lebanon. He kept apologizing for  his not paying frequent visits, as the wont with typically Lebanese villagers, because his brother’s wife, Amal, was still undergoing excruciating  larynx cancer treatment in Beirut. Amal is only 19, has a son, and her husband is only 20. The villagers call him ‘walad’, kiddo,  because he got married at a young age. The doctor says that the cancer has spread in most parts of her larynx and that ostracizing the malign cells without killing the healthy ones is getting harder by the day. The doctor says, “ Don’t let your hopes high”.  The listeners hearken at the story of Amal’s ordeals of the invasive, malignant cancer with compassion but not with surprise. In this in-land Lebanese village, and many, many others in Lebanon, cancer incidences are ubiquitous. During  my short summer stay, I incessantly hear about or know of the death of a man, the diagnosis of a woman, or t...