Monday, December 26, 2022

More educational malpractice: the sad legacy of Everyday Math, II

Here's a follow-up post to my first "sad legacy of Everyday Math" post, in which I concluded by saying that

You can’t blame the mathematical deficiencies of these 4th and 5th graders on their parents: both the private school and the after school program select for parents who care about education. You can’t blame it on the kids: my kids, who clearly wanted to learn, had been admitted [to our after school program] in part based on their behavior.

Picking up from there, the second post proceeded as follows:

You also can't blame it on language problems; these kids are fluent in English. In fact, there's really only one thing outside the Everyday Math curriculum that one can possibly point a finger to, and that is that these immigrant parents (many of them don't speak English) don't realize what many native-born parents already know: namely, that they can't count on the schools to fully educate their children.

So these kids are a case study in what happens when you leave math instruction entirely up to Everyday Math practitioners. And the answers to this question are slowly coming in.

For several of the 5th graders I work with, it turns out that not only do they not know how to borrow across multiple digits; they also don't know their basic addition and subtraction "facts." In other words, they don't automatically know that, say 5 plus 7 is 12, or that 15 - 8 is 7; instead they count on their fingers.

This got me thinking about addition and subtraction "facts." Back in my day, there was no issue of kids  learning these facts as such. Yes, we memorized our multiplication tables. But we never set about deliberately memorizing that 5 plus 7 is 12. Why? Because the frequency of the much-maligned "rote" calculations we did ensured that we, in today's lingo, constructed this knowledge on our own.

Back in my day, a typical third grade arithmetic sheet looked something like this:




And a typical fourth grade arithmetic sheet looked something like this:





But in Reform Math programs like Everyday Math, such pages filled with calculations are only occasional, and each problem involves a much shorter series of calculations. Here's a set from 4th grade Everyday Math:





Each multi-digit addition problem amounts to a series of simple addition problems. For example, adding two two-digit numbers involves adding at least two pairs of numbers; three if one is regrouping. Adding three three-digit numbers can involve 8 iterations of simple addition. Some of the problems in the second traditional math sheet involve as many as 17 iterations of simple addition.

In the traditional 4th grade math scenario, we may have had 25 problems per day like those in the first two sheets above, 5 days a week. With Everyday Math, you might get, at best, 25 problems like those in the second two sheets above per week.

Putting it all together, the resulting difference in the amount of practice with basic addition "facts" is quite large.  5 (days) times 25 (problems) times (say, as an average of iterations of simple addition) 10 for the traditional math curriculum versus 1 (day) times 25 (problems) times (average iterations) 3 for the Everyday Math curriculum. Assuming I'm not screwing up my arithmetic, that's 1,250 vs. 75 basic addition calculations per week. No wonder so many of those who are educated exclusively through Everyday Math don't know their "addition facts" by grade 5!

Ah, but surely their "conceptual understanding" is deeper. Note the calls for "ballpark estimate" at the bottom of each Everyday Math problem, where traditional math simply has you calculate. Stay tuned: in my next post on this topic, I'll discuss the state of conceptual understanding in my Everyday Math mal-educated 5th graders.

Thursday, December 22, 2022

Educational malpractice for the sake of Reform Math

Jo Boaler has been in the news once again, this time as one of the authors of the controversial new California Math framework

So I thought I'd resurrect this old post, inspired by, and based on part on, an email message I received back in 2013 from James Milgram, an emeritus Professor of Mathematics at Stanford University.

Professor Milgram is known in the education world for his comprehensive critique of a study done by Jo Boaler, an education professor at Stanford, and Megan Staples, then an education professor at Purdue. Boaler and Staples' paper, preprinted in 2005 and published in 2008, is entitled Transforming Students’ Lives through an Equitable Mathematics Approach: The Case of Railside School. Focusing on three California schools, it compares cohorts of students who used either a traditional algebra curriculum, or the Reform Math algebra curriculum The College Preparatory Mathematics (CPM). According to Boaler and Staple's paper, the Reform Math cohort achieved substantially greater mathematical success than the traditional math cohorts.

In early 2005 a high ranking official from the U.S. Department of Education asked Professor Milgram to evaluate Boaler and Staples' study. The reason for her request? She was concerned that, if Boaler and Staples' conclusions were correct, the U.S. department of education would be obliged, in Milgram's words, "to begin to reconsider much if not all of what they were doing in mathematics education." This would entail an even stronger push by the U.S. educational establishment to implement the Constructivist Reform Math curricula throughout K12 education.

Milgram's evaluation of Boaler and Staples' study resulted in a paper, co-authored with mathematician Wayne Bishop and statistician Paul Clopton, entitled A close examination of Jo Boaler's Railside Report. The paper was accepted for publication in peer-reviewed journal Education Next, but statements made to Milgram by some of his math education colleagues caused him to become concerned that the paper's publication would, in Milgram's words, make it "impossible for me to work with the community of math educators in this country"--involved as he then was in a number of other math education-related projects. Milgram instead posted the paper to his 
Stanford website.

This past October a bullet-point response to Milgram's paper, entitled "When Academic Disagreement Becomes Harassment and Persecution," appeared on Boaler's Stanford website. A month ago, Milgram posted his response and alerted me to it. I have his permission to share parts of it here.

Entitled Private Data - The Real Story: A Huge Problem with Education Research, this second paper reviews Milgram et al's earlier critiques and adds several compelling updates. Together, the two papers make a series of highly significant points, all of them backed up with transparent references to data of the sort that Boaler and Staple's own paper completely lacks.

Indeed, among Milgram et al's points is precisely this lack of transparency. Boaler and Staples refuse to divulge their data, in particular data regarding which schools they studied, claiming that agreements with the schools and FERPA (Family Educational Rights and Privacy Act) rules disallow this. But FERPA only involves protecting the school records of individual students; not those of whole schools. More importantly, refusals to divulge such data violate the federal Freedom of Information Act. Boaler's refusal also violates the policies of Stanford University, specifically its stated "commitment to openness in research" and its prohibitions of secrecy, "including limitations on publishability of results."

Second, Milgram et al's examination of the actual data, once they were able to track it down via California's education records, shows that it was distorted in multiple ways.

1. Boaler and Staple's chosen cohorts aren't comparable:

It appears, from state data, that the cohort at Railside [the pseudonym of the Reform Math school] was comprised of students in the top half of the class in mathematics. For Greendale, it appears that the students were grouped between the 35th and 70th percentiles, and that the students at Hilltop were grouped between the 40th and 80th percentiles. [Excerpted from Milgram; boldface mine]

2. Boaler and Staple's testing instruments are flawed:

Our analysis shows that they contain numerous mathematical errors, even more serious imprecisions, and also that the two most important post-tests were at least 3 years below their expected grade levels.  [Excerpted from Milgram; boldface mine]

3. The data comparing test scores on California's standardized tests (STAR) comes from a comparison of test scores from students not involved in Boaler and Staple's study:

The students in the cohorts Boaler was studying should have been in 11th grade, not ninth in 2003! So [this] is not data for the population studied in [Boaler and Staple's paper]. This 2003 ninth grade algebra data is the only time where the Railside students clearly outperformed the students at the other two schools during this period. There is a possibility that they picked the unique data that might strengthen their assertions, rather than make use of the data relevant to their treatment groups.   [Excerpted from Milgram; boldface mine]

4. The most relevant actual data yields the opposite conclusion about the Reform Math cohort's mathematical success relative that of the traditional math cohorts:

  • The most telling data we find is that the mathematics remediation rate for the cohort of Railside students that Boaler was following who entered the California State University system was 61%
  • This was much higher than the state average of 37%
  • Greendale's remediation rate was 35% o and Hilltop's was 29%

5. School officials at "Railside" report that the results of the reform math curriculum are even worse than Milgram et al had originally indicated:

A high official in the district where Railside is located called and updated me on the situation there in May, 2010. One of that person's remarks is especially relevant. It was stated that as bad as [Milgram et al's original paper] indicated the situation was at Railside, the school district's internal data actually showed it was even worse. Consequently, they had to step in and change the math curriculum at Railside to a more traditional approach.

Changing the curriculum seems to have had some effect. This year (2012) there was a very large (27 point) increase in Railside's API score and an even larger (28 point) increase for socioeconomically disadvantaged students, where the target had been 7 points in each case.

6. Boaler’s responses to Milgram et al provide no substantiated refutations of any of their key points

In response to comments on an article on Boaler's critique of Milgram, Boaler states:

"I see in some of the comments people criticizing me for not addressing the detailed criticisms from Milgram/Bishop. I am more than happy to this. [...] I will write my detailed response today and post it to my site."

However, as Milgram notes in his December paper:

As I write this, nearly two months have passed since Boaler's rebuttal was promised, but it has not appeared. Nor is it likely to. The basic reason is that there is every reason to believe [Milgram et al's paper] is not only accurate but, in fact, understates the situation at "Railside" from 2000 - 2005.

In a nutshell: under the mantle of purported FERPA protection, we have hidden and distorted data supporting a continued revolution in K12 math education--a revolution that actual data show to be resulting, among other things, in substantially increased mathematics remediation rates among college students. Ever lower mathematical preparedness; ever greater college debt. Just what our country needs.

Nor is Boaler's Reform Math-supporting "research" unique in its lack of transparency, in its lack of independent verification, and in its unwarranted impact on K12 math practices. As Milgram notes,

This seems to be a very common occurrence within education circles.

For example, the results of a number of papers with enormous effects on curriculum and teaching, such as [Diane Briars and Lauren Resnick's paper "Standards, assessments -- and what else? The essential elements of Standards-based school improvement"] and [J. Riordan and P. Noyce's paper, "The impact of two standards-based mathematics curricula on student achievement in Massachusetts"] have never been independently verified.

Yet, [Briars and Resnick's paper] was the only independent research that demonstrated significant positive results for the Everyday Math program for a number of years. During this period district curriculum developers relied on [Briars and Resnick's paper] to justify choosing the program, and, today, EM is used by almost 20% of our students. Likewise [Riordan and Noyce's paper] was the only research accepted by [the U.S. Department of Education's] What Works Clearinghouse in their initial reports that showed positive effects for the elementary school program ``Investigations in Number, Data, and Space,'' which today is used by almost 10% of our students.

As Milgram notes:

Between one quarter and 30% of our elementary school students is a huge data set. Consequently, if these programs were capable of significantly improving our K-12 student outcomes, we would surely have seen evidence by now.

And to pretend that such evidence exists when it doesn't is nothing short of educational malpractice.

Monday, December 19, 2022

The sad legacy of everyday math

[Everyday Math, I gather, is still very much in use, and so I thought it worthwhile to recycle this old post.]

Twice this past week I saw shocking examples of the cumulative effects of Everyday Math. Last Thursday I visited a nearby private school with sliding scale tuition and a diversity of students. For years the school had used Everyday Math, but recently, with the encouragement of a friend and colleague of mine who advises schools on math curricula, they’d begun to use Singapore Math. They’re phasing it in gradually, however, and currently don’t introduce it until 4th grade. For the first few grades, like nearly every other school in Philadelphia, they use Everyday Math.

So the 4th graders I observed had only been using Singapore Math since September. Their teacher was walking them through a topic in the 3rd grade Singapore Math curriculum: how to multiply and reduce fractions. And no one in the class who tried to answer the teacher’s questions got a single answer right. They didn’t know how to find ¼ of a 20, and they didn’t know how to reduce 5/20.

The next day I spent my first session of the school year with a group of children of French African immigrant parents who had enrolled them in an after school enrichment program I’m involved with. They were four Everyday Math-educated 5th graders, and I was exploring their mastery of addition and subtraction. Addition went fine: they know how to stack numbers and carry from one digit to the next. Subtraction was another story.

Heartened by their success adding two three-digit numbers, I asked them how to do 1000 - 91. All but one of the five students were stumped. Most got the same number: 1011. Two things had stumped them: 0 - 1, which they thought was 1, and how to borrow across more than one digit. So I gave them an easier problem, 100-71--and they were equally stumped, again getting answers that were larger than the number they were subtracting from. So I began the tricky process of teaching them how to borrow across more than one digit.

The great thing is that they were hooked. When I asked them whether their answers should be bigger or smaller than the number they were subtracting from, they all answered “smaller.” When I then asked them whether their answers were, in fact, smaller, they looked down at their sheets, and then up at me, and I had their undivided attention. These are good kids: they want to learn. And they like math.

You can’t blame the mathematical deficiencies of these 4th and 5th graders on their parents: both the private school and the after school program select for parents who care about education. You can’t blame it on the kids: my kids, who clearly wanted to learn, and had been admitted in part based on their behavior; in the private school classroom I saw, they were very well behaved. You can’t blame it on class size: the classes I observed contained between 7 and 12 students. You can’t blame it on the teachers: the teachers I saw seemed well above average in their ability to engage their students in the material at hand.

No, I’m afraid there’s only one thing we can blame here, much as the developers of the Everyday Math monolith would like to claim otherwise.

Friday, December 16, 2022

Where calculators don't help

How much longer would it have taken for the flight to take off if one of the younger flight attendants had taken charge?

Another resurrected post:

Everyday arithmetic: where calculators don't help

What with the proliferating meme that calculators can substitute for most real-world human calculations, and with restaurant bills that increasingly calculate the various tip possibilities for you (15%, 20%, 25%), it's harder and harder to find examples of cases in which, say, there's no good substitute for knowing your multiplication tables.

But on a recent airplane flight, one such case suddenly jumped out at me. Before the flight took off, the flight attendants were asked to verify that all the passengers were on board. There were six seats per row on what seemed to be a maxed-out flight, and one of the older flight attendants knew exactly what to do. Walking down the aisle, she rapidly counted by 6's: 6, 12, 18, 24, etc...

Think about it... There's no faster way to count rows of airplane passengers than to apply your memorized multiplication tables.

How much longer would it have taken for the flight to take off, I couldn't help wondering, if one of the younger flight attendants had taken charge?

Tuesday, December 13, 2022

False choices in remediation: "addition and subtraction over and over again" or Marxism and Shakespeare

I suspect the rejection of true remediation has only grown over the years since I wrote this, as schools, bowing to the Common Core standards, increasingly expect nearly all students to engage with the same material based not on academic readiness, but on what year and month they happened to be born in.


Another false choice in remediation: "addition and subtraction over and over again" or Marxism and Shakespeare

A recent New York Times Education Supplement article entitled "Rigorous SchoolsPut College Dreams Into Practice" showcases Bard College's new "early college high school" for disadvantaged students in Newark.

For the uninitiated (like yours truly until I read this), an early college high school is one that merges high school courses with "some college." As the article explains:

Students can earn both a high school diploma and an associate degree, and some are set on the path to a four-year degree.

The early college high school is also a growing movement:

There are now more than 400 early college high schools across the country — North Carolina has 76 of them — educating an estimated 100,000 students.

Across the country in communities like Newark, the early college high school model is being lauded as a way to provide low-income students with a road map to and through college.

What makes early college high schools different from, say, a college prep track of a regular high school? For one thing, they seem to be specifically geared at students who need "catching up," and they aim to offer an alternative to remediation:

The ethos of early college high schools: catch students up, not by relegating them to the kind of remedial classes required at community colleges but by bombarding them with challenging work. At the Bard school, that means works by Dante, Locke and W.E.B. Du Bois that have populated and enriched the lives of their more affluent peers.

Can Dante really replace traditional remediation? At Bard's Newark branch:

Students say the transition has been tough. Al-Nisa Amin, now a sophomore, remembers slumping over a math problem that first year, crying out of sheer frustration. But she has stuck it out, partly because she is scared of being sent to a zoned high school.

In particular, the article cites students who were getting A's at their zoned high schools now getting D's and F's. And here's what their Shakespeare seminars are like:

Flipping through their Signet Classic paperbacks and scribbling notes, they reviewed the first act of “Twelfth Night,” intuitively understanding that Orsino, Duke of Illyria, had become obsessed with Olivia. When their professor, David Cutts, asked what was going on in Orsino’s heart, several called out matter-of-factly: “Love.” The class then discussed the vagaries of love at first sight, and voted on whether they believed in it. Most didn’t.

Some were confused by the shipwrecked noblewoman Viola and her motives in disguising herself as a servant. “He’s rich, so why is she trying to hide?” one student asked, befuddled. Another hypothesized: “I think she’s interested in him.”

Then there's the Marxism and Postmodernism seminar:

...which on this day involved mulling over a densely written essay by the Marxist political theorist Fredric Jameson on the meaning of self in a postmodern world. Reading aloud a 2006 article in The Economist titled “Post-Modernism Is the New Black,” one student stumbled over “facade,” “anachronistic” and “grandeur” — words that would seem fair game for late high school.

Another student wanted to know: “What’s a phenomenon?” One inquired about the meaning of “sinister.”

The article cited Auschwitz as an example of how the Enlightenment had “given birth” to totalitarianism. Not one of the 10 students knew what Auschwitz was. Debate ensued over whether it was a city in Switzerland, Russia or Poland. Their professor finally interjected: “It’s usually used as the big example of the Holocaust.”

It's important to note that this may not be representative of early college high school seminars in general:

In similar classes in Bard’s New York schools, students’ vocabulary, communication skills and historical knowledge appear noticeably more advanced.

But, as the article notes:

The disparity raises an uncomfortable question. Can students who are so behind be brought up to college level in a few efficient years, even with good teachers and good intentions?

So far, the school has lost 7 of the 36 students who entered in 2011 as first-year college students and 20 of the 87 who entered as high school freshmen.

Najee has repeated one class. Both Miles and Billy have repeated several. More than half the class had to repeat one of the required seminars in a monthlong intensive at the end of the last school year.

The article proceeds to elaborate on the pedagogical philosophy of these early college high schools:

Taken as a group, early college high schools place a premium on teaching rudimentary study skills — how to take notes, how to interact with professors, where the best spot is to sit in a classroom. But the greatest emphasis is on thinking. Students are encouraged to see themselves as participants in an academic world, and as interested in gaining knowledge as in getting good grades. Dr. Ween calls it “joining the debate.”

So far, so good. But, as the next paragraph makes clear, "thinking" and "joining the debate" mean something troubling specific and Constructivist:

The students at Duplin Early College High School in eastern North Carolina take an applied math class in which they learn about velocity and graphing by building roller coasters out of wire, piping and masking tape. Then they are asked to defend the project. At the Dayton Early College Academy in Ohio, students learn about constitutional law in mock trials. And at Bard, in an environmental science class, students read articles about the effect biofuel is having on corn prices and debate the merits of renewable energy.

These activities do not address the deficits that colleges most often need to remediate: those in essay writing and pre-calculus.

“You cannot pull off an early college high school successfully without fundamentally changing pedagogy,” said Joel Vargas, vice president of Jobs for the Future, a nonprofit organization based in Boston that develops early college high schools. He calls it the opposite of “chalk and talk.”

I agree with Vargas' first statement. But fundamentally changing pedadogy in a way that prepares disadvantaged kids for high school means less project-based learning and more direct instruction and pen and paper exercises ("chalk and talk"?).

As articles like this one make clear, the edworld is increasingly convinced that remediation can be bypassed:

Gone is the thinking that students must master all the basics before taking on more challenging work.

“Traditionally, what has happened is that kids who come in below standards are put in a remedial track and they do addition and subtraction over and over again,” said Cecilia Cunningham, executive director of the Middle College National Consortium, a network of more than 30 early college high schools. “They’re bored out of their minds and the message is: ‘You really can’t do this.’ ”

To maintain beliefs like these, the edworld depends on such false dichotomies. Remediation = doing addition and subtraction over and over again = boring kids out of their minds and telling them they're incapable. The only alternative is project-based learning and Postmodernism seminars. It simply doesn't occur to them that remediating a child's academic skills at their Zone of Proximal Development is less likely to bore them and make them feel incapable than forcing them through a seminar on the meaning of self in a postmodern world (even with the implicit threat of having to return to their local high schools if they don't play ball).

Do such tactics nonetheless work? In general:

Studies show that high school students who take classes in which they get both college and high school credit — often referred to as dual-credit courses — fare better academically.

A study last year of more than 30,000 Texas high school graduates found that those who took college-level classes in high school were more likely to have finished college after six years.

Studies like these, however,

aren’t able to determine if it is the type of students drawn to college-level coursework that makes the difference. And no long-term studies have been conducted about early college high school students and college graduation. [Emphasis mine]

A refreshing voice in the edworld wilderness comes from emeritus professor Sandra Stotsky

who notes that there is not any substantial evidence that the model being tried out in Newark will help at-risk students get through four years of college. Dr. Stotsky finds the idea that students should have to go to college to get a good high school education counterintuitive, and has called on educators to refocus their efforts on making high school coursework more challenging.

Otherwise, the opposite may happen--at the college level:

Critics also worry about rushing students through the material and pushing them prematurely onto college campuses, thus dumbing down classes for the other students.

It's interesting that, for all the permeation of the "learning differences" and "differentiated instruction" memes, what current trends in education have most amounted to in practice is a one-size-fits all curriculum in which the same hands-on, arts-based "multiple learning styles" assignments are inflicted on everyone and neither advancement nor remediation are allowed.


Thursday, December 8, 2022

Is there really no Theory of Mind deficit in autism? Part II: the validity of the standard Theory of Mind measures

Cross-posted at FacilitatedCommunication.org.

In my previous post on Gernsbacher and Yergeau’s 2019 paper, Empirical Failures of the Claim That Autistic People Lack a Theory of Mind, I discussed problems with the authors’ arguments that the original studies that showed Theory of Mind deficits in autism have failed to replicate and been overturned by later studies. As the article continues, the authors embark on a second line of argumentation—this one concerning the inherent validity of the various ToM tests.

Morton A. Gernsbacher, Professor of Psychology, University of Wisconsin

The various ToM tests, Gernsbacher and Yergeau argue, fail to “converge” (that is, agree in their conclusions) on anything that meaningfully distinguishes autism.  The ToM tests they consider are:

  • The Strange Stories Test, which measures the ability to deduce the social/emotional reasons for characters’ behaviors in a narrative sequence

  • The Animated Triangles Test, which measures the ability to ascribe emotions to self-propelled, interacting shapes in a short animation

  • The Eyes Test, which measures the ability to recognize emotions from facial information that is restricted to a rectangular region around the eyes

  • The Faux Pas Test, which measures the ability to detect social blunders

  • The false-belief tests, which measures the ability to make inferences about the beliefs held by (or about) individuals who are missing key pieces of information

    • The Sally-Anne “unexpected location change”  test, in which Sally doesn’t witness Anne changing the location of her marble

    • The Smarties “unexpected contents” test, in which a candy box contains a pencil rather than candy

This “lack of convergent validity among theory-of-mind tasks,” they conclude, “undermines the core construct validity of theory of mind.”

Gernsbacher and Yergeau cite, in particular, the repeated failure of the Strange Stories test to converge with the Eyes test or the Animated Triangles test; and the repeated failure of the Eyes test to converge with the Faux Pas test. These test pairs, however, measure quite different social phenomena: detecting emotion from eyes; making sense of social behavior in a narrative; mentalizing about animated shapes; and detecting social blunders.  Thus, none of these convergence failures is particularly surprising. The most related of these two phenomena, the Strange Stories and the Faux Pas tests, are one pair whose degree of convergence Gernsbacher and Yergeau do not mention. The lack of convergence of the other pairs may only mean that other variables, like individual variation within the autism spectrum in frequency of eye contact or in verbal skills, also play a role. This doesn’t rule out that each of these tests tap into an underlying social factor that is atypical in autism.

But Gernsbacher and Yergeau go on to make an even more surprising claim: that even within the false-belief tests there is lack of convergence, with different false-belief tasks failing to correlate significantly with one another. The only specific examples of non-convergent false-belief tasks in the studies they cite, however, are (1) first- vs. second-order false-belief tasks and (2) one study in which the Smarties task fails to correlate with the Sally-Anne test. As far as first vs. second-order false-belief tests go, we’ve already seen their non-convergence: first-order false-belief tests correlate with autistic traits (and language development); second-order false-belief tests correlate with working memory/processing abilities.  But that does not undermine the validity of first-order false-belief tests in singling out autism.

As far as the Smarties vs. the Sally-Anne test, one study, Hughes (1998), found that performance on the Smarties test was not correlated with the Sally-Anne test. Nonetheless, each of these tests correlated with a different ToM task. The Sally-Anne test, which has to do with the location of hidden objects, correlated with a penny-hiding task, while the Smarties test, which has to do with a deceptive container, correlated with a deception task. Furthermore, as Lind and Bowler (2009) point out, the Sally-Anne task involves a narrative and a relatively simple test question (“Where will Sally look for her marble?”), while the Smarties task does not involve a narrative but does involve a more complex test prompt (“Billy hasn’t seen the box. When he comes in I’ll show him the box just like this and ask what’s in there. What will he say?”). Different skillsets (comprehension of hiding vs. deception; following a narrative vs. parsing a complex question), therefore, may be involved in passing each one, even if each also taps into the same underlying ToM skills. Beyond this one study, studies in general (e.g., Ozonoff, 1991, mentioned *in my earlier post*, for Smarties; or Kerr and Durkin, 2004, mentioned below, for Sally-Anne) have found both of these tests to be disproportionately difficult for individuals with autism.

Gernsbacher and Yergeau’s other two references for lack of convergence among false-belief tests do not involve autistic subjects. Charman & Campbell (1997) addresses individuals with learning disabilities; Duval et al. (2011), individuals with age-related cognitive decline.

In short, to the extent that there is a lack of convergence between ToM tests as applied to autistic individuals, this lack of convergence does not undermine the tests’ validity as measures that tap into the various aspects of the ToM deficits in autism.

REFERENCES:

Charman, T., & Campbell, A. (1997). Reliability of theory of mind task performance by individuals with a learning disability: A research note. Journal of Child Psychology and Psychiatry, and Allied Disciplines, 38, 725–730. https://doi.org/10.1111/j.1469-7610.1997.tb01699.x

Duval, C., Piolino, P., Bejanin, A., Eustache, F., & Desgranges, B. (2011). Age effects on different components of theory of mind. Consciousness and Cognition, 20, 627–642. https://doi.org/10.1016/j.concog.2010.10.025

Gernsbacher, M. A., & Yergeau, M. (2019). Empirical Failures of the Claim That Autistic People Lack a Theory of Mind. Archives of scientific psychology7(1), 102–118. https://doi.org/10.1037/arc0000067

Hughes, C. (1998). Executive function in preschoolers: Links with theory of mind and verbal ability. British Journal of Developmental Psychology, 16, 233–253. https://doi.org/10.1111/j.2044-835X.1998.tb00921.x

Kerr, S., & Durkin, K. (2004). Understanding of thought bubbles as mental representations in children with autism: Implications for theory of mind. Journal of Autism and Developmental Disorders, 34, 637–648. https://doi.org/10.1007/s10803-004-5285-z

Lind, S. E., & Bowler, D. M. (2009). Language and theory of mind in autism spectrum disorder: the relationship between complement syntax and false belief task performance. Journal of autism and developmental disorders39(6), 929–937. https://doi.org/10.1007/s10803-009-0702-y

Ozonoff, S., Rogers, S. J., & Pennington, B. F. (1991). Asperger’s syndrome: Evidence of an empirical distinction from high-functioning autism. Journal of Child Psychology and Psychiatry, and Allied Disciplines, 32, 1107–1122. https://doi.org/10.1111/j.1469-7610.1991.tb00352.x

Monday, December 5, 2022

A linguist's meditation on youth, age, and the passage of time

Does the passage of time--forward or backward--make things younger or older?

A short time ago, the youngest countries (e.g. Eritrea and South Sudan) came into existence. 

A long time ago, America was one of the youngest countries. 

As time moves forward, America gets older and older and so do we.

Going back in time, as we get younger and younger (and eventually don't exist), we eventually find older and older languages, cultures, civilizations, and historical figures... 

But long ago, many of these things were new...

Saturday, December 3, 2022

Shakespeare for everyone?

A recent class discussion on full inclusion for autistic students led me and my students to the Shakespeare Question. This question is forced upon us by the fact that the Common Core Standards, for all their general vagueness about curricula, require nearly all high school students to read Shakespeare, Only the most intellectually disabled are exempted. 

But plenty of non-to-minimally disabled students can't handle the Bard, even with page-by-page glossaries. Some teachers send students over to SparkNotes or No Fear Shakespeare. And then there's...

21st century Shakespeare

[An Out in Left Field post from 2011]

One thing that got lost in the discussion following the NY Times' recent exposé of a school district in Arizona that has spent $33 million on laptops, big interactive screens, and so-called educational software and is asking for $46.3 million for more of the same even as its test scores are falling and its music, art, and phys ed classes are being cut, is what is happening to Shakespeare:

Amy Furman, a seventh-grade English teacher here, roams among 31 students sitting at their desks or in clumps on the floor. They’re studying Shakespeare’s “As You Like It” — but not in any traditional way.

In this technology-centric classroom, students are bent over laptops, some blogging or building Facebook pages from the perspective of Shakespeare’s characters. One student compiles a song list from the Internet, picking a tune by the rapper Kanye West to express the emotions of Shakespeare’s lovelorn Silvius.

This is a great way to ruin literature in general, but reducing Shakespeare in particular to the psycho/sociological perspectives of his characters is particularly ruinous.  Despite claims by some that Shakespeare invented the human, human psychology was not Shakespeare's strong suit. His characters are constantly falling in love--or tumbling into murderous jealousy--at the drop of a hat or handkerchief; switching love interests at the lifting of a disguise; and ridiculously prone to deception (c.f., e.g., the "transvestite comedies"), persuasion ( And be it moon, or sun, or what you please: An if you please to call it a rush-candle, Henceforth I vow it shall be so for me.) or (most infamously) forgiveness.

No, what Shakespeare excelled in was--of course--language, and the radiant gems of wisdom into which he worked it. Some of my favorites:

Uneasy lies the head that wears a crown.

 or:

The undiscovere'd country, from whose bourn
No traveller returns 

or:

Be not afraid of greatness: some are born great, some achieve greatness, and some have greatness thrust upon them.  

What would these look like as status updates on Facebook?

Wednesday, November 30, 2022

Celebrating creativity?

Creativity is one of those things we say we celebrate--along with honesty, open-mindedness, critical thinking, neurodiversity, and diversity of opinion. But creativity also strikes me as one of those areas in which everyone has an agenda.

Here's another post from 2011.

The virtues of creativity

In this blog I've repeatedly complained about wrong-headed priorities involving creativity, explaining answers, learning styles, self-esteem, and relevance. In the next few blog posts, I'm going to argue that each of these things should still play some role K12 education. I begin, today, with creativity.

As regular readers of the blog know, I've repeatedly criticized the practice of grading students on creativity when there is no established protocol for teaching it or for measuring it objectively (beyond the crude, reductionist, and pointless tallying of things like "illustrations per page" and "colors per illustration"). Does that mean that creativity should play no role in K12 education? Not at all. Teachers can, and should, attempt to inspire creativity in their students and to be creative themselves in their teaching.

But inspiring creativity does not mean simply exhorting students to "be creative." Far more promising is helping students appreciate the creative works of others--in anything from poems, to stories, to essays, to paintings, to clever solutions to math problems. Or giving them inspiring assignments or prompts--e.g., extend such and such a bizarre opening sentence into the first of a story; or extend such and such a random squiggle into a page-sized illustration. Or taking them on exotic outings--topiary gardens; whirligig exhibits; kitchen instrument concerts; or La Compagnie Transe Express.

Teachers themselves can be creative, too, even in the driest of subjects, and without reducing the subject matter to meaningless mush. A friend of mine who teaches at St Ann's in Brooklyn--a school whose combination of a classical education and a creative teaching staff is perhaps unparalleled--dresses up every Wednesday like a puritanical schoolmarm, topped with a puritan cap and equipped with a dunce cap, teaches her 5th graders the fundamentals of grammar and style. They love it, and they learn it.

Ironically, at the same time that students are increasingly graded on "creativity," schools are inspiring it less than ever. They're increasingly more likely to punish than to reward teachers who don't adhere to their scripted lines. And, their language-policed, all-about-me curriculum, which prefers mirrors to windows, provides way too few windows, in particular, into creativity at its most inspiring.

 

Perhaps that's why America is less and less a land of innovation--except for first-generation immigrants, assuming they don't go back home.

Monday, November 28, 2022

Campaigning vs. governing, nth edition

The world would be a better place if successful campaigning and successful governing didn’t depend on such different skill sets.

Here’s an updated version of an old post

Narcissistic Leadership

Personality, of course, is a multi-dimension affair, and there are as many personality spectra as there are personality dimensions. But one possible spectrum I've been wondering about goes something like this:

Autism___________Asperger's__________Normal Range_________________________Narcissism

At both ends you see varieties of extreme self-absorption. In autistic spectrum disorders, the source and the essence are socio-cognitive: core to autism is difficulty inferring, automatically and accurately, the perspectives of other people. In narcissism, the source and essence are emotional: self-love interferes both with gut-level empathy and with the desire to consider others' perspectives. Source and essence aside, perspective taking deficits obviously impair pro-social behavior, including, for example, the ability to lead others.

But because narcissists are better at hiding (or “masking”) their disorder, it tends to be only those on the autistic side of the spectrum who strike employers as bad candidates for leadership positions. Indeed, because narcissists also exude confidence and charm, many of us consider them leadership whizzes. Wouldn't it be nice if a study cited in last week's weekend 
Wall Street Journal could make us think twice:

When narcissists get assigned to leadership roles, they impress their charges with authority and confidence. They also underperform, a new study finds.
...

Workers assigned to narcissistic leaders tended to report that they were authoritative and effective. But, in fact, narcissists-led groups shared information less effectively than the others and picked the wrong candidate more often.

The study offers a cautionary tale to businesses, the authors said: Narcissists are equipped to ace job interviews, for example, but are unlikely to live up to expectations.

At some level, this explains a lot of what's wrong in the world. But why does it not surprise me that many people who ace job interviews are "unlikely to live up to expectations"?

On the flipside, I suspect that many on the other end of the spectrum, while floundering in job interviews, would make great employees, if only someone gave them the chance.


Friday, November 25, 2022

The Rectangle: an appreciation

[A recently rediscovered post from 11 years ago]

Only the Triangle gets a course to itself, but, as a tool for teaching math concepts in general, the Rectangle (including the special case of the Square) is, as it were, unparalleled.  Rectangles are, of course, the basic constituent of Singaporean bar modeling, substituting for x and y in word problems that might otherwise require algebra:

Mr. Lim read 10 pages from his book on Monday. Mr. Smith read three times as many pages as Mr. Lim. Their friend, Mr. Samy, read 15 pages more than Mr. Smith. How many pages did Mr. Samy read from his book?

Rectangles are also a good way to visualize multiplication:

5 × 9:



74 × 368:


including the commutative law of multiplication:

And the distributive law:

 a (b  + c) = ab + bc:

And equivalent fractions:

And the factoring or multiplying out of quadratics:


And the area of triangles:

And the area under a curve:


Indeed, so versatile is the unassuming rectangle that I'm sure I've missed a number of other areas for which it is a powerful conceptual building block.


Wednesday, November 23, 2022

Is there really no Theory of Mind deficit in autism? Part I: is it all about language instead?

Cross-posted at FacilitatedCommunication.org.

In this post and five subsequent posts, I review the final article in my series on Morton Gernsbacher’s FC-friendly articles on the nature of autism. In this last article, Empirical Failures of the Claim That Autistic People Lack a Theory of Mind, Gernsbacher and Yergeau (2019) go further than any of Gernsbacher’s previous articles in making the case against autism as a socio-perceptual, socio-cognitive disorder. In particular,  Gernsbacher and Yergeau claim that the original studies that showed Theory of Mind deficits in autism have failed to replicate and been overturned by later studies.

Morton A. Gernsbacher, Professor of Psychology, University of Wisconsin

Gernsbacher and Yergeau open with a straw-man characterization of what Theory of Mind proponents purportedly have said: “The assertion that autistic people lack a theory of mind—that they fail to understand that other people have a mind or that they themselves have a mind—pervades psychology”. The more common claim, rather, is that autism involves some degree of deficit in the recognition/awareness of emotions in other people (e.g., attending to and recognizing facial expressions, or deducing specific emotions from behavior), combined with some degree of deficit in socio-cognitive perspective taking: in figuring out how to respond appropriately to another person’s emotional needs; in deducing the belief set of another person when those beliefs conflict with one’s own.

Of course, Gernsbacher and Yergeau’s argument depends, to some extent, on what exactly is meant by Theory of Mind (henceforth, ToM). To some extent, ToM has traditionally been defined by the long-standing tests used to assess ToM: false-belief tests like Sally-Anne test and the Smarties test; Happé’s Strange Stories; Baron-Cohen’s Mind in the Eyes test (which we’ll call “Eyes test” for short); and the Animated Triangles test. Here is a quick summary of the key tests:

  • The Strange Stories Test measures the ability to deduce the social/emotional reasons for characters’ behaviors in a narrative sequence.

  • The Eyes Test measures the ability to recognize emotions from facial information that is restricted to a rectangular region around the eyes.

  • The Animated Triangles Test measures the ability to ascribe emotions to self-propelled, interacting shapes in a short animation.

  •  The false-belief tests measure the ability to make inferences about the beliefs held by (or about) individuals who are missing key pieces of information. Two common examples are:

    • The Sally-Anne “unexpected location change”  test, in which Sally doesn’t witness Anne moving her marble from a basket to a box

    • The Smarties “unexpected contents” test, in which a candy box contains a pencil rather than candy

False-belief tasks are the most cognitively complex of the ToM tasks: they measure what we can call “cognitive perspective taking”—the ability to put oneself in someone else’s cognitive shoes and deduce their state of knowledge or beliefs. Gernsbacher and Yergeau begin their article by reiterating one of the arguments made in Gernsbacher and Frymiare (2005) about how failure in such tests correlates with language development rather than with autism. Typically developing children need to reach a certain level of language development to pass false-belief tests, and children with non-autism related language delays—e.g., deafness—are also delayed in passing false-belief tests. But Gernsbacher and Frymiare (2005), as I discuss in my *earlier post*, do not address the fact that autistic individuals need to reach a much higher language level than other groups do to pass these tests. This fact suggests that those autistic individuals who do pass false-belief tests may be doing so through an atypical mechanism—i.e., a more deliberative reasoning out, or “hacking out” through language.

Gernsbacher and Yergeau, on the other hand, do attempt to address the “hacking out” hypothesis. Their counterargument is that it is hard to reconcile the “hacking out through language” strategy with the language impairments common to autism. They ask: “How and why would autistic people preferentially use language to ‘hack out’ the answers while non-autistic people, without communication impairments, do not?” The answer, of course, is that language, given the challenges of autism, may nonetheless still be autistic people’s best option—assuming, of course, that they reach the requisite level of linguistic mastery. Moreover, the fact that the level of linguistic mastery required for autistic people to pass false-belief tests is considerably higher than that required for non-autistic individuals suggests that, when autistic individuals pass false-belief tests, language is doing more work—i.e., playing a bigger role—than it does for non-autistic people.

Gernsbacher and Yergeau also reiterate Gernsbacher and Frymiare’s (2005) arguments about the complex language involved in false-belief tests—and how, therefore, failing to pass false-belief tasks may merely reflect language difficulties. However, as I noted in reference to the 2005 paper, there are non-verbal versions of the false-belief tests—also called “implicit” false-belief tests—that rely only on anticipatory gazing (e.g., towards where Sally will look for her marble) as opposed to answering a complex question (e.g., “Where does Sally think her marble is”). One implicit false-belief experiment is discussed in Senju et al. (2009). They find that the autistic participants, unlike the non-autistic participants, failed to look anticipatorily at the location where Sally would look for her marble. Indeed, this is even the case with high functioning adults with Asperger’s who (as would be expected from their verbal mental ages) were able to the standard (verbal) ToM tests, including the Sally-Anne task and a second order ToM test called the “ice cream test”.

Instead of discussing these eye-tracking versions of the false-belief tests, Gernsbacher and Frymiare cite, as their only example of nonverbal false-belief tasks, the picture-sequencing experiment of Baron-Cohen et al. (1986). This experiment compared autistic and non-autistic performance on sequencing pictures into three kinds of sequences: mechanical sequences (e.g., a balloon rising into the sky and then being punctured by a tree branch),  behavioral sequences (e.g., a child entering a store and exchanging money for merchandise), and “intentional”/ToM sequences (e.g., a child putting some candy in the box, leaving the room, and then, after someone else opens the box and eats the candy, returning, re-opening the box, and looking disappointed). Regarding these findings, Gernsbacher and Yergeau say:

Four research teams, of whom we are aware, have published attempts to directly replicate these results—and none could do so. Using the same stimuli, procedures, and analyses, no other research team has replicated the finding that autistic participants perform significantly worse than typically developing participants on the “intentional” picture sequences.

The studies cited by Gernsbacher and Yergeau here are Ozonoff et al. (1991); Oswald and Ollendick (1989); Buitelaar et al. (1999), and Brent, Rios et al. (2004). And while all of them do indeed fail to replicate Baron-Cohen et al.’s results, some of their other findings do suggest that ToM deficits nonetheless characterize autism.

For example, Ozonoff et al. found “significantly poorer emotion perception abilities in this sample of autistic individuals.” They also found that a subset of the autistic group (as opposed to the Asperger’s group) performed significantly less well on the first order ToM test (the Smarties test) and that this performance correlated with the emotion perception scores.

Brent et al., meanwhile, found that the autistic participants “performed worse than typically developing controls on the mentalizing Strange Stories task and the Eyes task. Their conclusion: “children with ASD are impaired relative to typically developing children in advanced theory of mind abilities.”

 As some of these authors acknowledge, the picture sequencing tasks may not accurately capture ToM deficits. This is not the same thing as finding that ToM deficits do not exist.

Despite Ozonoff’s replications of relative difficulty for autistic people on first-order false-belief tests, Gernsbacher and Yergeau claim that “Baron-Cohen, Leslie, and Frith (1985)’s seminal study reporting that autistic participants are prone to fail first-order False Belief tasks… is also prone to fail replication.” Here they cite Dahlgren & Trillingsgaard (1996);  Russell & Hill (2001); Oswald & Ollendick (1989); Fitzpatrick et. al. (2013); Yirmiya & Shulman (1996); Yirmiya et. al. (1998); Moran et. al., (2011).

To the extent that these studies fail to replicate an autism-specific tendency to fail false-belief tasks, it is largely because of the linguistic confounds discussed above: all children fail the standard false-belief tests before they achieve a certain language level, and all children pass the standard false-belief tests above a certain language level. What’s specific to autism, as we discussed, is the need to achieve a significantly higher level of linguistic functioning than in other groups in order to pass explicit false-belief tests. Furthermore, Buitelaar et. al. (1999), cited elsewhere in Gernsbacher and Yergeau’s piece, found that performance on first-order false-belief tests correlated with autistic traits as measured by a standard measure of autism symptomology (the CARS).

Gernsbacher and Yergeau then turn to second-order false-belief tasks. These are tasks that involve calculating what one character will believe about another character’s (faulty) belief (“Where does Anne think that Sally will look for her marble?”). Here they point to studies showing that autistic participants are no more prone to fail second-order false-belief tasks than other populations are—a fact that is indeed borne out by the various studies. However, there are some key things to keep in mind. First, since passing a first-order false-belief test is a prerequisite for passing a second-order false-belief test, these studies only include individuals who pass first-order false-belief-tests. In autism, as we discussed, passing first-order false-belief tests means having achieved a significantly higher verbal mental age than in other groups. And what the studies (e.g., Tager-Flusberg & Sullivan, 1994) propose is that, at this point, challenges other than autism-specific challenges like mentalizing figure most importantly. Prime candidates are executive functioning (EF) challenges like working memory and cognitive load/processing demands.

In addition, the ability to pass second-order false-belief tests does not rule out ToM difficulties in the real world. For example, Leekam & Prior (1994), despite finding that the autistic participants were no more prone than other groups to fail second-order false-belief tests and could “make appropriate social judgements about lies and jokes”, also found difficulties, as per parent reports, in real-world ToM skills. These reports turned up difficulty both with lying, in particular white lies, and with joking—and here there was no difficulty between second-order “passers” and “failers.” Overall, Leekam & Prior’s result support that “theory of mind development is delayed relative to the mental age of normal children”.

Similarly, Bauminger & Kasari (1999) report that, despite the similar performance by autistic and non-autistic participants on second-order false-belief tests, “the autistic children, as a group, gave more irrelevant or wrong responses when asked to explain their answers to the belief question.” (Such explanations are not generally part of standard false-belief tests). Such performance, they note, suggests the ability to pass second-order false-belief tests does not rule out deficits in social understanding. They cite similar findings by Ozonoff & Miller (1995) and others. As Bauminger & Kasari put it, “Even very able children with autism, or children with Asperger syndrome, have difficulties in applying theory of mind when faced with a real social situation.”

One theory of mind task that is intended to measure real-world skills are Happé’s Strange Stories—stories that involve real-world mentalizing situations. But here, too, Gernsbacher and Yergeau argue that there has been a failure to replicate—a failure, specifically, to replicate Happé’s (1994) finding that autistic participants who pass first- or second-order False Belief tests nonetheless fail the Strange Stories test.

But once again there is the confounding role of language. The Strange Stories, by their very nature, are grounded in language and involve much more verbiage than false-belief tests. Even if the syntax isn’t as complex, there are a greater range of vocabulary words. Consistent with this, Sheeren et al. found that age and verbal abilities (as well as general reasoning abilities), not autism diagnosis, were the determining factors for success with Strange Stories. Consistent, in turn, with Sheeren et. al., those studies that found no deficiencies with Strange Stories involved adults with high functioning autism or Asperger’s (Senju et. al., 2009; Schuwerk et. al., 2018; Ponnet et. al., 2004).

As for younger children, Gillotte et. al., (2004) report that, despite similarities in overall performance between autistic and non-autistic groups, the former gave significantly more inappropriate mental state answers (e.g., angry instead of sad) to the Strange Stories questions. This suggests, Gillotte et al. propose, that many children with autism, despite “severe ToM difficulties”, had “learned to apply generalized compensatory strategies such as offering an explanation in terms of mental states to any question as to why a person acted in a particular way.”

Another study involving younger children, cited by Gernsbacher and Yergeau as failing to replicate Happé (1994)’s Strange Stories results, is White, et. al. (2009). While children with autism who showed ToM impairment on other tests performed “significantly more poorly” than controls on the Strange Stories tests, they also performed more poorly on counterparts in which the stories were based on animals. Perhaps this suggested to Gernsbacher and Yergeau, inasmuch as animals don’t have minds, that the issue was something other than ToM. White et al.’s explanation, rather, is that “a mentalizing [ToM] deficit may affect understanding of biologic agents even when this does not explicitly require understanding others’ mental states.”

Returning, now, to autistic adults, even those highly verbal autistic participants who performed well on the Strange Stories tests struggled with other ToM measures. Schuwerk et. al. (2018), for example, report that such participants scored significantly lower on the Eyes test and also showed less accurate anticipatory looking in an eye-tracking-based implicit false-belief test akin to the one discussed above. Schneider et. al. (2013) also found that, despite similar performance on explicit ToM tests, autistic individuals showed no evidence of anticipatory looking (“tracking”) in an implicit false-belief test, “even over a one-hour period and many trials.” They conclude that “the systems involved in implicit and explicit ToM are distinct” and that “impaired implicit false-belief tracking may play an important role in ASD”.

Furthermore, Murray et. al. (2017) show that, when the Strange Stories are substituted with an animated counterpart in which language is less of a confounding factor, the differences between autistic and non-autistic performances increase, with the former showing more relative difficulty. As Murray et. al. note,

Adults with ASD had lower scores, indicating difficulties with social cognition that could not be explained by general cognitive factors (e.g., verbal ability) and were specific to understanding the intentions behind nonliteral language in communication.

Ponnet et. al., (2004)’s findings were similar. On the standard “static” versions of both the Strange Stories and Eyes tests, the adults with Aspergers performed similarly to typically developing adults. But in the more naturalistic empathic accuracy task, in which participants attempted to infer a person’s thoughts and feelings in a videotape of that person in a naturally occurring conversation with another person, the individuals with Asperger’s performed significantly worse. Roeyers et. al., (2001) obtained similar results in similar experiments with adults with pervasive developmental disorder (PDD).  As they put it:

These findings suggest that the mind-reading deficit of a subgroup of able adults with PDD may only be apparent when a sufficiently complex naturalistic assessment method is being used.

One study cited by Gernsbacher and Yergeau as failing to replicate Happé (1994)’s Strange Stories results, Spek et. al. (2010), in fact did replicate it. They found that HFA and Aspergers adults “were impaired in performance of the Strange Stories test and the Faux-pas test [which taps the ability to detect social blunders] and reported more theory of mind problems than the neurotypical adults.” It was only their performance on the Eyes test—which, as we noted earlier, tests the ability to infer specific emotions from the eye-region of faces—that their performance was equivalent to the other groups.

As for the Eyes test, Gernsbacher and Yergeau note, correctly, that it involves sophisticated vocabulary (words like “envious” and “fantasizing”)—though we should add that word definitions are often provided as part of the test. Gernsbacher and Yergeau then cite research, both on autistic individuals and on individuals with non-autistic language challenges, that shows that “the best predictor of Reading-the-Mind-in-the-Eyes” is vocabulary. But one of these studies still shows connections, if weaker, to cognitive empathy and emotion perception (Olderbak et. al., 2015). In addition, as we’ve discussed earlier, vocabulary itself is grounded in social engagement: early vocabulary development is a function of frequency of responses to Joint attention (RJA). Given how subsequent vocabulary builds on earlier vocabulary, it could be that vocabulary levels, even late in development, correlate with early (and even current) social engagement—the more so with words like “envious” and “fantasizing”. If so, then there may be an underlying social variable that underlies both vocabulary, especially socio-emotional vocabulary, and performance on the Eyes test.

This blog post has gone on long enough. For those who’ve made it this far, here’s a quick recap:

  • Some ToM tests use language as a testing medium, and thus require a certain threshold level of linguistic comprehension. This can result in language skills being a bigger causal factor in test performance than mentalizing/perspective-taking skills are—except when we consider the causal role played by mentalizing/perspective-taking skills in the acquisition of those language skills.

  • Autistic individuals need to acquire significantly more language than non-autistic individuals do before they can pass false-belief tests. This suggests that autistic individuals are using language to reason through these tests in a more deliberate way than their non-autistic counterparts do.

  • Second-order false belief tests (“Where does Ann think that Sally will look for her marble?”) appear to be primarily measuring something other than social reasoning skills (e.g., working memory and cognitive load/processing demands). But second-order false belief tests are generally administered only to those who pass first-order false-belief tests—those, in order words, who have already reached a significant social-cognitive milestone.

  • Second-order false belief tests aside, the patterns of performance on ToM tests observed in autistic vs. non-autistic individuals, including in the articles cited by Gernsbacher and Yergeau, consistently indicate ToM difficulties in autism.

  • The most impaired ToM skills are those involving

    • automatic (as opposed to deliberate) perspective-taking (e.g., automatic gaze shifting to where Sally will look for her marble)

    • socially naturalistic settings.


REFERENCES

Bauminger, N., & Kasari, C. (1999). Brief report: theory of mind in high-functioning children with autism. Journal of autism and developmental disorders29(1), 81–86. https://doi.org/10.1023/a:1025974701090

Baron-Cohen, S., Leslie, A. M., & Frith, U. (1985). Does the autistic child have a "theory of mind"?. Cognition21(1), 37–46. https://doi.org/10.1016/0010-0277(85)90022-8

Baron-Cohen, S., Leslie, A. M., & Frith, U. (1986). Mechanical, behavioural and Intentional understanding of picture stories in autistic children. British Journal of Developmental Psychology, 4, 113–125. https://doi.org/10.1111/j.2044-835X.1986.tb01003.x

Brent, E., Rios, P., Happé, F., & Charman, T. (2004). Performance of children with autism spectrum disorder on advanced theory of mind tasks. Autism : the international journal of research and practice8(3), 283–299 (p. 286). https://doi.org/10.1177/1362361304045217

Buitelaar, J. K., van der Wees, M., Swaab-Barneveld, H., & van der Gaag, R. J. (1999). Theory of mind and emotion-recognition functioning in autistic spectrum disorders and in psychiatric control and normal children. Development and psychopathology11(1), 39–58. https://doi.org/10.1017/s0954579499001947

Dahlgren, S. O., & Trillingsgaard, A. (1996). Theory of mind in non-retarded children with autism and Asperger’s syndrome: A research note. Journal of Child Psychology and Psychiatry, and Allied Disciplines, 37, 759–763. https://doi.org/10.1111/j.1469-7610.1996.tb01469.x

Fitzpatrick, P., Diorio, R., Richardson, M. J., & Schmidt, R. C. (2013). Dynamical methods for evaluating the time-dependent unfolding of social coordination in children with autism. Frontiers in Integrative Neuroscience, 7, Article 21. https://doi.org/10.3389/fnint.2013.00021

Gernsbacher, M. A., & Frymiare, J. L. (2005). Does the Autistic Brain Lack Core Modules?. The journal of developmental and learning disorders9, 3–16.

Gernsbacher, M. A., & Yergeau, M. (2019). Empirical Failures of the Claim That Autistic People Lack a Theory of Mind. Archives of scientific psychology7(1), 102–118. https://doi.org/10.1037/arc0000067

Gillott, A., Furniss, F., & Walter, A. (2004). Theory of mind ability in children with specific language impairment. Child Language Teaching and Therapy, 20, 1–11. https://doi.org/10.1191/0265659004ct260oa

Happé, F. G. E. (1994). An advanced test of theory of mind: Understanding of story characters’ thoughts and feelings by able autistic, mentally handicapped, and normal children and adults. Journal of Autism and Developmental Disorders, 24, 129–154. https://doi.org/10.1007/BF02172093

Leekam, S. R., & Prior, M. (1994). Can autistic children distinguish lies from jokes? A second look at second-order belief attribution. Journal of Child Psychology and Psychiatry, and Allied Disciplines, 35, 901–915. https://doi.org/10.1111/j.1469-7610.1994.tb02301.x

Moran, J. M., Young, L. L., Saxe, R., Lee, S. M., O’Young, D., Mavros, P. L., & Gabrieli, J. D. (2011). Impaired theory of mind for moral judgment in high-functioning autism. Proceedings of the National Academy of Sciences of the United States of America, 108, 2688–2692. https://doi.org/10.1073/pnas.1011734108

Murray, K., Johnston, K., Cunnane, H., Kerr, C., Spain, D., Gillan, N., . . . Happé, F. (2017). A new test of advanced theory of mind: The “Strange Stories Film Task” captures social processing differences in adults with autism spectrum disorders. Autism Research, 10, 1120–1132. https://doi.org/10.1002/aur.1744

Olderbak, S., Wilhelm, O., Olaru, G., Geiger, M., Brenneman, M. W., & Roberts, R. D. (2015). A psychometric analysis of the reading the mind in the eyes test: Toward a brief form for research and applied settings. Frontiers in Psychology, 6, Article 1503. https://doi.org/10.3389/fpsyg.2015.01503

Oswald, D. P., & Ollendick, T. H. (1989). Role taking and social competence in autism and mental retardation. Journal of Autism and Developmental Disorders, 19, 119–127. https://doi.org/10.1007/BF02212723

Ozonoff, S., & Miller, J. N. (1995). Teaching theory of mind: a new approach to social skills training for individuals with autism. Journal of autism and developmental disorders25(4), 415–433. https://doi.org/10.1007/BF02179376

Ozonoff, S., Rogers, S. J., & Pennington, B. F. (1991). Asperger’s syndrome: Evidence of an empirical distinction from high-functioning autism. Journal of Child Psychology and Psychiatry, and Allied Disciplines, 32, 1107–1122. https://doi.org/10.1111/j.1469-7610.1991.tb00352.x

Ponnet, K., Buysse, A., Roeyers, H., & De Corte, K. (2005). Empathic accuracy in adults with a pervasive developmental disorder during an unstructured conversation with a typically developing stranger. Journal of Autism and Developmental Disorders, 35, 585–600. https://doi.org/10.1007/s10803-005-0003-z

Roeyers, H., Buysse, A., Ponnet, K., & Pichal, B. (2001). Advancing advanced mind-reading tests: Empathic accuracy in adults with a pervasive developmental disorder. Journal of Child Psychology and Psychiatry, and Allied Disciplines, 42, 271–278. https://doi.org/10.1111/1469-7610.00718

Russell, J., & Hill, E. L. (2001). Action-monitoring and intention reporting in children with autism. Journal of Child Psychology and Psychiatry, and Allied Disciplines, 42, 317–328. https://doi.org/10.1111/1469-7610.00725

Schneider, D., Slaughter, V. P., Bayliss, A. P., & Dux, P. E. (2013). A temporally sustained implicit theory of mind deficit in autism spectrum disorders. Cognition, 129, 410–417. https://doi.org/10.1016/j.cognition.2013.08.004

Schuwerk, T., Priewasser, B, Sodian, B, & Perner, J. (2018). The robustness and generalizability of findings on spontaneous false belief sensitivity: A replication attempt. Royal Society Open Science, 5, Article 72273. https://doi.org/10.1098/rsos.172273

Senju, A., Southgate, V., White, S., & Frith, U. (2009). Mindblind eyes: An absence of spontaneous theory of mind in Asperger syndrome. Science, 325, 883–885. https://doi.org/10.1126/science.1176170

Spek, A. A., Scholte, E. M., & Van Berckelaer-Onnes, I. A. (2010). Theory of mind in adults with HFA and Asperger syndrome. Journal of Autism and Developmental Disorders, 40, 280–289. https://doi.org/10.1007/s10803-009-0860-y

Tager-Flusberg, H., & Sullivan, K. (1994). A second look at second-order belief attribution in autism. Journal of Autism and Developmental Disorders, 24, 577–586. https://doi.org/10.1007/BF02172139

White, S., Hill, E., Happé, F., & Frith, U. (2009). Revisiting the strange stories: Revealing mentalizing impairments in autism. Child Development, 80, 1097–1117. https://doi.org/10.1111/j.1467-8624.2009.01319.x

Yirmiya, N., Erel, O., Shaked, M., & Solomonica-Levi, D. (1998). Meta-analyses comparing theory of mind abilities of individuals with autism, individuals with mental retardation, and normally developing individuals. Psychological Bulletin, 124, 283–307. https://doi.org/10.1037/0033-2909.124.3.283

Yirmiya, N., & Shulman, C. (1996). Seriation, conservation, and theory of mind abilities in individuals with autism, individuals with mental retardation, and normally developing children. Child Development, 67, 2045–2059. https://doi.org/10.2307/1131608