Neuromyths in Education: Why do they persist?

In a post last year we discussed issues with the lack of evidence-based education, and during some recent professional development sessions and conversations at my university, this came to mind again, as some of the ideas that seem to be taken for granted in higher education seem to have very little supporting evidence. Indeed, education does seem an area where some of the so called “neuromyths” persist, and are even championed.

I came up against this when challenging some of the constructivists and postmodern educational ideas being discussed by a PhD student and a senior member of faculty in education. I was told “I didn’t realize people actually still thought that sort of thing” as if I was some sort of ludite dinosaur. Sadly, I have found members of nursing faculty are often patronized in such ways by academics from other disciplines, which usually signals to me an inability to make any useful counter-argument to a point, and possibly also a rather closed mind!

Neuromyths are really ideas about neurological/cognitive processes that have been repeated often enough to become considered as fact. Unfortunately some misconceptions about the brain persist in the classroom and beyond. Let us consider a few of these established ideas that pervade higher education that have mainly arisen from dubious educational psychology and persist as contemporary wisdom.

Left Brain – Right Brain

The idea is often suggested that people are predominately left or right brained in terms of their skills and aptitudes. E.g. left-brain predominant = logical and mathematically skilled, more organized and systematic whilst right brain predominant = artistic and creative. Just google “left-brain right brain” for many examples. Current research suggests regardless of personality or skill set, you use both the right and left hemispheres of your brain together to perform everyday tasks. Although certain functions, such as speech production, handedness, and facial recognition, tend to be dominated by one side of the brain in the great majority of people, most tasks require parallel processing from both hemispheres. The integration of input is made possible by the fibre connections between right and left sides of the brain called the corpus callosum. Unless an entire hemisphere is completely removed or damaged, no one should really be considered to be “right”- or “left”-brained

The Utility of Learning and Teaching Styles 

As educator James Atherton notes: most teachers would not argue with the proposition that people learn or teach in different ways. This has given rise to a whole host of theories of learning (and teaching) styles. There are at least 71 different learning styles inventories published. However, the assumptions of the “styles” adherents in education are that it is possible to develop a relatively simple typology of learning or teaching styles and then develop test instruments to ascertain where individuals fit, teach to address them, and (more worryingly)  assess the quality of teaching with reference to this.

The evidence to support this is unfortunately weak at best. The research does not support the notion that there are hard-wired styles, and many of the theories conflate learning styles with learning strategies, cognitive theories, or personality type theories. Certainly, students may well have learning preferences but they are not as clear-cut as these various inventories suggest, and motivation would appear to over-ride them every time (Pashler et al. 2008; Scott, 2010). Nevertheless, if you look at many university education and professional development sites they continue to be taught at sage wisdom, and many commercial enterprises exist who are happy to sell you a test.

The Learning Pyramid

The following diagram (or versions of it) appears in around 15,000 web sites (if you do a reverse image search on google – or simply search “learning pyramid”), and yet the evidence that supports it is very vague. It purportedly depicts the degree of retention of material with various teaching methods.









It may come from early work by Dale (1946/1969) but even the US based National Training Laboratories Institute for Applied Behavioural Science (who cite it) admit: “NTL believes it to be accurate” but says that they “can no longer trace the original research that supports the numbers” Magennis and Farrell (2005:48). It is also often conflated with the notion of the “cone of experience” in education and the Washington Post also did a nice article on the flaws with it in 2013. Again, there is probably some use and truth in the notion that some teaching methods will work better for some subjects and in some situations. However, the idea that there is a strong validated theoretical model with clearly defined categories is far from the truth.

Multiple Intelligences and Thinking Hats

Howard Gardner’s multiple intelligences model and Edward de Bono’s thinking hats are other good examples of theories I often hear discussed or quoted to support pedagogic approaches. Yet both are also good examples of modern neuromyths. Gardner first proposed his theory of different types of intelligence in 1983. Since then, it has undergone incremental changes, including the addition of one additional intelligence (bringing the total to eight). These different forms of intelligence have been advocated as a basis for changing the way in which we teach. But, repeated research and meta-analysis has found no evidence that individuals actually conform to Gardner’s theoretical categories. Also, according to a 2006 study many of Gardner’s “intelligences” correlate with the g factor, supporting the idea of a single dominant type of intelligence. 

Indeed, even intelligence quotient (IQ) theory itself is commonly misinterpreted. The first IQ test was made by French psychologist Alfred Binet in 1905, and since then the IQ test has become the most recognized tool for predicting academic and professional success. However, although well validated as a psychometric measure there are a number of myths about it that persist, such as:

  • It measures intelligence
  • IQ cant change
  • IQ is genetic

Lastly as a predictive factor for success, it would seem rather simplistic, and although generally a good predicator of performance, does not explain the many confounding examples of successful people who have lower IQ scores than those less successful.

The Thinking Hats site reads like rather a satire on the subject. It was originally proposed by Edward de Bono in 1985. The premise of the method is that the human brain thinks in a number of distinct ways which can be deliberately challenged, and hence planned for use in a structured way allowing one to develop tactics for thinking about particular issues. However,there is virtually no empirical evidence supporting the model, and it has often been parodied.

In the end, Gardner’s theory or de Bono’s thinking hats interesting ideas but probably not all that helpful for adoption in formal education.

 You Only Use 10% of Your Brain

Again this seems a a widespread common belief, but though the 10-percent myth is widespread, recent neuro-imaging technology has conclusively destroyed this. While not all of the brain is active all at once, functional magnetic resonance images (fMRI) show several brain areas are at work for any given activity, depending on what function is needed, and that we use the majority of our brain matter daily.

Lack of Theoretical Development and Testing?

Overall, I fear part of the problem here is the trend towards accepting postmodern constructivist epistemologies, over thorough scientific investigation, or what I might call the “its all good” syndrome. I worry that this ambivalence towards good evidence in academic inquiry is actually gathering steam, rather than diminishing with key examples being the current rise of so-called integrative science and quackademic medicine. Good scientific practice involves developing ideas into theories, and testing them repeatedly to identify the best of a set of competing hypotheses or explanations. That does not mean we have found the truth but the best explanation given our current understanding. An approach that accepts them all as equally valid explanations of the world offers little in practical value, apart from the ongoing generation of even more unsubstantiated theory.

Enough Already!

The call that we need more research into these theories is often suggested, but we should also recognize the comes a point when it is reasonable to say we have enough evidence, and move on to something new. It is not so much that these neuromyths are wrong, but that the evidence base and/or research methodology is flaky at best, and they have often been misinterpreted and generalized beyond their legitimate use, and make little sense in the real world of education. So, time to move educational theory on towards more productive areas where student performance can actually be shown to improve, such as with the use of improved feedback/formative assessment strategies.

There is an excellent balanced chapter on “neuromyths” from a recent book by the co-ordinator of the site Howard-Jones, that is well worth a look.

Onwards and Upwards



Atherton J. Read more on misrepresentation, myths and misleading ideas on James Atherton’s site at:

Dale, E. (1969) Audiovisual methods in teaching, third edition.  New York: The Dryden Press; Holt, Rinehart and Winston.

Doidge, N. The Brain That Changes Itself: Stories of Personal Triumph from the Frontiers of Brain Science. Penguin Books, 2007

Howard-Jones P (2009) Introducing Neuroeducational Research London; Routledge.

Jarrett C. Why the Left-Brain Right-Brain Myth Will Probably Never DiePsychology Today, June 27, 2012

Magennis S and Farrell A (2005) “Teaching and Learning Activities: expanding the repertoire to support student learning” in G O’Neill, S Moore and B McMullin Emerging Issues in the Practice of University Learning and Teaching, Dublin; All Ireland Society for Higher Education/Higher Education Authority

Pashler H, McDaniel M, Rohrer D and Bjork R (2008) “Learning Styles; concepts and evidence” Psychological Science in the Public Interest vol. 9 no.3; available on-line at accessed 21 May 2014.

Scott, C. (2010) The Enduring Appeal of ‘Learning Styles’ Australian Journal of Education 2010 54: 5 DOI: 10.1177/000494411005400102

Visser, Beth A.; Ashton, Michael C.; Vernon, Philip A. (2006), “g and the measurement of Multiple Intelligences: A response to Gardner”, Intelligence 34 (5): 507–510,


Do You Understand Lupine Ways of Knowing? The value of reductio ad absurdum in scientific debate.

This week I thought I would raise the rather contentious  issue of the reductio ad absurdum argument (also known as argumentum ad absurdum). This is the ancient form of logical argument that seeks to demonstrate that an argument or idea is nonsense by showing that a false, ludicrous, absurd result follows from its acceptance, or alternatively that  an argument is sound as a false, untenable, or absurd result follows from its denial.

The nature of this argument has venerable roots and it is well documented as a form of logic in ancient Greece, used by such luminaries as Xenophranes, Socrates, and Plato . However, in modern academia there seem to be rather polarized views on it. 1) that it trivializes an argument and belittles the person taking a particular position or 2) that is is a valid and reasonable way of demonstrating that an idea is unsound. There also seems to be a cultural aspect in that I have found it used more frequently in Europe, whereas in North America it is somewhat frowned upon in many academic circles.

Naturally, as Rog and I are somewhat subversive and agitative academics (I use the term loosly) we are in full support of it, and to this end have just published a paper in Nursing Inquiry using exactly this form of argument to challenge the established wisdom of a specific postmodern argument for alternative ways of knowing. This paper was based on an earlier blogpost on this very blog site. Here, we use the ad absurdum argument to note that the principles used to support Carper’s  four ways of knowing can equally well be used to support a more creative typology (in this case including, arcane knowing, and lupine knowing).

Naturally, as with any form of intellectual rationale the argument is only as good as the fundamental data and facts it is based upon. Therefore, an ad absurdium argument can be misused, or poorly constructed. It is also often used erroneously as a Straw Man argument.

Considering what is absurd and what isn’t is a tricky thing for anyone, and particularly problematic in science.  For example, many Victorian scientists scoffed at the thought of powered flight, and even Einstein had issues with the notion of black-holes. Therefore, identifying absurdity is not something easily undertaken, as it may simply be the ideas presented are highly original or unconventional. The bacteria Helicobacter Pylori being suggested as a cause of gastric ulceration is a good example, as this theory was not readily accepted by the medical community for several years, despite good evidence.

Also, this is not the same as absurdity as used in common parlance. Commonly absurd positions are seen as ridiculous, or foolhardy, but an argument ad absurdum does not suggest the person making the argument should be ridiculed or lampooned. After all, we have all believed ridiculous things at one time or another; for western children the notion that Santa Clause brings all the children in the world toys on one night a year is a case in point! For the purpose of scientific thinking, for something to be demonstrated as absurd here we really need to see that there is inconsistency in the arguments presented. An absurd position may be considered one that is contrary to reason, irrational, or ludicrous to follow due to the practical implications of believing it. Unfortunately, several concepts now accepted and used in modern science arose in exactly this fashion: Quantum physics for example. However, repeated scientific observation and empirical data have proved quantum theory correct. So, paradigms change with time and we should be cautious about suggesting any position is ridiculous.

From a pragmatic position, I would argue an argument that can be demonstrated as fallacious by analysing its components, and demonstrating inconsistencies, or that you can demonstrate by accepting it you are also supporting associated positions that make no sense and have no practical value, then an ad absurdum position can be used effectively to demonstrate these weaknesses.

At the end of the day the sensitivities invoked by this form of argument are worth considering, and it is a form of rationale that is not easy to develop effectively. However, as long as the use of it involves demonstrating the nonsense an idea or position presents, rather than attacking the person making the argument, I would suggest it is a useful form of analysis. As a scientist if you are prepared to make any case, hypothesis or argument, you should be prepared to have it challenged and debated, and defend it. If the position is sound it will survive this critique, and win through. That is what good science is all about, but to make sound ad absurdum arguments you have to have a good working knowledge of the logical fallacies to start with.  They can also be a lot of fun too, and if this form was good enough for Socrates…



Carper B.A. (1978), “Fundamental Patterns of Knowing in Nursing”, Advances in Nursing Science 1(1), 13–24

Garrett B.M. & Cutting R.L. (2014) Ways of knowing: realism, non-realism, nominalism and a typology revisited with a counter perspective for nursing science. Nursing Inquiry. Retrieved 21 May 2014.

Rescher N. (2009) Reductio ad absurdumThe Internet Encyclopedia of Philosophy. Retrieved 21 May 2014.




True Grit: Just how gritty are modern students?

Hi folks,

Apologies for the silence on the blog for the last few weeks. I have been down at a Scholarship of Teaching and Learning  Conference in Georgia last week (note to self: never fly United again, its like a budget bus trip), and Roger has been publishing his new creative science teaching textbook (of which more will be revealed shortly). Also he has somehow lost the ability to login and post here. Anyhow, we should have that sorted shortly. So, enough apologies and a quick blog post for now, and we shall do our best to catch up in April.

At the conference I came across an interesting paper by Dr. Ted Cross from Grand Canyon University exploring the use of the Grit psychometric scale and correlations with academic success in online doctoral students. In case you have not come across it, the Grit personality trait measurement is a 5 point positive, non-cognitive scale based on an individual’s passion for a particular long-term goal or end state coupled with their motivation to achieve this objective. The maximum score on the scale is 5 (extremely gritty), and  the lowest scale on this scale is 1 (not at all gritty). Generally Grit can be simply defined as perseverance and passion for long-term goals (Duckworth et al., 2007). There is an interesting TED talk by Angela Lee Duckworth explaining her work on the scale here:

Overall it seems a well validated psychometric tool, and unlike IQ which remains controversial for any practical predictive ability, and can be culturally insensitive; Grit appears to have less of these problems. Also, Grit, unlike many traditional measures of performance is not tied to intelligence. Interestingly, Cross reported Grit was well correlated with GPA, but not with Standard Test scores in the USA. Unfortunately, he found the scale was not sensitive enough to be useful to explain differences within the doctoral students he tested it with, or predict likely success (after all these are already a highly-skewed, high achieving population and the test is designed for general population use).

To test your own grit try the Penn State University online test here, or the paper one here:  I came out as more gritty than 75% of Americans, which seems perfectly reasonable, given my career history, but penchant to always want to explore some new aspect in my work.

So what does this mean for us scientists? Well Grit is also associated with longer term and multi-year goals and science is a long-term business, requiring dedicated persistence to advance in a field, often in small steps. It is also worth considering how today’s millennial, multi-tasking, short attention span learners are likely to fare in careers in science, or if we can actually have any impact on this trait as educators, or make our educational practices more appealing to intelligent students who are less gritty.

Also, I must admit I wondered, given the range of papers presented at the conference (from the very postmodern to the very post-positivist) of the possibilities of developing a trait scale to predict the likelihood of a person being susceptible to believing in unscientific nonsense (such as homeopathy, ghouls and ghosts, bigfoot, alien abductions, etc etc.). Maybe there is one out there in the world of psychometric testing? Not so much a gullibility scale, as a tendency towards fallibility scale. This would seem a personality trait that could be measured with practical uses. Knowing your own tendency towards fallibility could help in the analytical process. After all, even the most intelligent individual can be fooled quite easily (just look at famous scientists and academics who adopt bizarre and unfounded beliefs later in their careers: Dr Oz springs to mind amongst others). I must investigate further!

Onwards and upwards,



Duckworth, A.L., Peterson, C., Matthews, M.D., & Kelly, D.R. (2007). “Grit: Perseverance and passion for long-term goals”. Personality Processes and Individual Differences, 92 (6), p. 1087.