Author Archives: ryansmerek

When Do We Change Our Minds? Think of a Jenga Tower

I recently read Philip Tetlock’s excellent book Superforecasting: The art and science of prediction. (1) It is about improving our ability to make forecasts, which includes updating our beliefs as we learn new information about how the future may unfold.

There is one section of the book that discusses belief updating and Tetlock (along with his co-author Dan Gardner) use a helpful metaphor for when we will or will not change our minds when encountered with new information. They say:

“Commitment can come in many forms, but a useful way to think of it is to visualize the children’s game Jenga, which starts with building blocks stacked one on top of another to form a little tower. Players take turns removing building blocks until someone removes the block that topples the tower. Our beliefs about ourselves and the world are built on each other in a Jenga-like fashion” (p. 162).

With this metaphor, they discuss how, if you are forecasting within your specialty, you can be more reluctant to discard certain beliefs when the domain is tightly-connected with your identity and self-worth.

I like the metaphor of beliefs as a Jenga Tower because it’s easy to be pessimistic that we’ll never change our minds when confronted with information that conflicts with our beliefs. In political science, they even find a strengthening of prior beliefs when confronted with conflicting evidence—called the “backfire” effect (Nyhan & Reifler, 2010). For example, if we are strong supporters of gun control or another hot-button issue, when we receive conflicting information that doesn’t support our preexisting beliefs, we’ll argue against it, thus strengthening our prior beliefs. This is certainly true in many circumstances, but occasionally we will update our views (even those close to the base of the Jenga tower) as “incongruency builds” over time (See Redlawsk, Civettini, & Emmerson, 2010). Yes, for many issues we won’t budge (at least in the short term), because it would bring the whole Jenga Tower crashing down, but for other beliefs near the top, we’ll slowly remove and update some of them in accordance with the evidence we are presented with.

So, the next time you seem to be in an intractable conflict, think about where you are in the Jenga Tower (and what this says about belief centrality). If you are near the base of someone’s self-worth and deeply-held identity, you have little chance of dislodging a core belief (at least in the moment). Instead, move up the Jenga Tower toward less foundational assumptions. For example, we can discuss whether tax credits for low-emission vehicles are a worthy policy to combat pollution, without having to agree and debate the causes of climate change (See Lewandowsky, Ecker, Seifert, Schwarz & Cook, 2012), This may be a more fruitful (and hopeful) way of conversing in domains that seem intractable, and eventually changing our (and others) minds.

(1) Don’t be put off by the heroically dramatic title. The book is highly-nuanced and impressively-grounded. I’d put it in the same category as Kahneman’s masterful Thinking Fast and Slow.


Lewandowsky, S., Ecker, U. K. H., Seifert, C. M., Schwarz, N., & Cook, J. (2012). Misinformation and its correction: Continued influence and successful debiasing. Psychological Science in the Public Interest, 13(3), 106-131.

Nyhan, B., & Reifler, J. (2010). When corrections fail: The persistence of political misperceptions. Political Behavior, 32, 303-330.

Redlawsk, D. P., Civettini, A. J. W., & Emmerson, K. M. (2010). The affective tipping point: Do motivated reasoners ever “get it”? Political Psychology, 31(4), 563-593.

Tetlock, P. E. & Gardner, D. (2015). Superforecasting: The art and science of prediction. New York: Crown Publishers.

Ryan Smerek is an assistant professor and assistant director of the MS in Learning and Organizational Change program at Northwestern University. He is the author of Organizational Learning and Performance: The Science and Practice of Building a Learning Culture.

Photo Source: Guma89/Wikimedia Commons

The Intuition of Truth

Whenever I finish a movie “based on a true story,” there comes a moment when the credits begin rolling and I want to know how much of it was true. This happened recently after finishing “The Founder,” about Ray Kroc and how he turned McDonald’s into an American institution. It led to a brief internet search to learn about the deal that was signed (or not signed) with the original McDonald’s brothers and whether the movie’s portrayal resembled reality. We only have so much time, of course, to dig into the “facts” of any movie “based on a true story,” but the point remains that we still have an intuitive reaction about the “truth.” We’ll certainly allow some “creative license,” but we still want the major points to largely resemble what occurred.

Likewise, I heard a talk by the author Tim O’Brien around the fall of 2010. He is the author, among others, of the book The Things They Carried, a collection of short stories about soldiers during the Vietnam War. I had read the book in the 1990s and was moved by the reflective tales about what war must have been like. During his talk, he mentioned an amusing story about his young son peeing into a waste paper basket in their family bathroom. I believe the waste paper basket was of the wicker-variety so that a pool of urine was all over the bathroom floor. I do not remember the exact reason his son did this, but the reason was humorous, along with the numerous other particulars of the story. The room was in laughter. When he finished, however, he said, “I just made that story up. It never happened.” He went on to say how it didn’t matter whether it happened or not, it was the emotional truth of the story. I can remember vehemently disagreeing. It did matter.

Both of these scenarios (movies based on a true story and Tim O’Brien’s “tale” of his son) sparked an intuitive reaction—that it matters whether something is true or not. But knowing “the truth,” is very complex, of course. There are many meanings of the word “truth” (Horwich, 2010, 2013). As Horwich (2013) describes, truth has been defined as “correspondence with the facts,” as “provability,” as “practical utility,” and as “stable consensus,” but “all turned out to be defective in one way or another—either circular or subject to counterexamples” (para. 12). Despite the contested nature of truth, we don’t want to “throw the baby out with the bathwater.”

This brings me to the excellent TED talk by Michael Patrick Lynch titled “How to see past your own perspective and find truth.” Lynch, a philosophy professor at the University of Connecticut, states:

“Skepticism about truth…is a bit of self-serving rationalization disguised as philosophy. It confuses the difficulty of being certain with the impossibility of truth. Look — of course it’s difficult to be certain about anything…But in practice, we do agree on all sorts of facts. We agree that bullets can kill people. We agree that you can’t flap your arms and fly. We agree — or we should — that there is an external reality and ignoring it can get you hurt.”

Thus, while truth has many meanings and it is difficult to be certain about anything, we still need “truth” as a regulatory ideal. At the very least, a belief in truth is functional in that it helps us get to the bottom of issues to develop a more solid footing for our beliefs and opinions. Asking of any claim “is it true?” helps spark critical thinking and dialogue to overcome a confirmatory stance to the world (1). So, we shouldn’t throw out the notion of truth with postmodern skepticism, even if it is difficult to ascertain, and the word, itself, means many different things. We should, at least, listen to our intuitive reaction and curiosity to understand if something is “true,” (i.e., did it actually occur), and not so easily abandon this ideal.

(1) See Dalio (2017) or Smerek (2017) for embedding the social norm of “pursuing truth” in an organizational setting.


Dalio, R. (2017). Principles: Life and work. New York: Simon & Schuster.

Horwich, P. (2010). Truth–meaning–reality. Oxford, UK: Oxford University Press.

Horwich, P. (2013, March 3). Was Wittgenstein right? The New York Times.

Lynch, M. P. (2017, April). How to see past your own perspective and find truth.

Smerek, R. E. (2017). Organizational learning and performance: The science and practice of building a learning culture. New York: Oxford University Press.

Ryan Smerek is an assistant professor and assistant director of the MS in Learning and Organizational Change program at Northwestern University. He is the author of Organizational Learning and Performance: The Science and Practice of Building a Learning Culture.

Photo Source: Dino Reichmuth/Unsplash



Forget Intelligence. Aim for Mental Complexity.

In reading the entire Oxford English Dictionary over the course of a year (a mere 20 volumes and 21,730 pages!), Ammon Shea was asked by NPR host Tom Ashbrook about some of the more compelling words he learned.

Shea mentioned “apricity,” which is the warmth of the sun in the winter. As he says:

“[Apricity] is a word that I sincerely hope I will never work into general conversation…I am not a fan of using big words for their own sake. However, I do find now that in the [wintertime] as I am bathed in that warmth of the winter sun, I am more cognizant of the fact that it’s happening, knowing there is a word with which to describe it” (Shea, 2008).

After learning the word apricity, Shea began to experience the world with more cognizance. You’ve likely had the same experience in learning a new concept about workplace dynamics. Perhaps you’ve learned about personality dimensions such as introversion-extraversion or how power dynamics can be at play in subtle ways. After learning these new ideas you experience workplace dynamics with more cognizance. You see more and experience more—just like Ammon Shea experiencing the warmth of the winter sun.

However, when trying to explain any kind of mental development, we often attribute that development to an increase in “intelligence.” Researchers in the field of adult development discuss a different form of intellectual competence—mental complexity (see Kegan & Lahey, 2009). Mental complexity is the variety of perspectives, concepts, and vocabulary we have to make sense of the world. With greater mental complexity, we can perceive more and take more effective action. We expand our action repertoire. As, “No one is ever free to do something he can’t think of” (Weick, 1979, p. 193).

Of course, there have been many critiques of intelligence that have tried to take it off its explanatory pedestal. Most notably is Keith Stanovich’s (2009) impressive work on the rational thinking habits that are dissociable from traditional measures of intelligence. In addition, Carol Dweck (2006), with her research on having a growth mindset, has helped us question our assumptions about intelligence. In particular, how our implicit belief about intelligence as a fixed quantity limits whether we seek challenges and persist, because challenges might reveal what is innately lacking. Instead, if we view the mind as a muscle, and capable of growing, we are more likely to pursue challenges to expand our “intelligence.” As important as this may be, we are still trapped within frame of intelligence.

If, however, we seek to expand our mental complexity, we can forgo any implicit assumptions we might have about intelligence. Instead, by aiming for mental complexity, we can more freely expand our cognizance of what we are experiencing and what is occurring in the organizations in which we work.

What does this mean you should do? Increase the number of educational actions you take: read more, persist through confusion to try and explain what you are experiencing with a more nuanced vocabulary. Do this—not to increase your intelligence—but to expand your mental complexity.


Dweck, C. S. (2006). Mindset: The new psychology of success. New York: Random House.

Kegan, R., & Lahey, L. L. (2009). Immunity to change: How to overcome it and unlock the potential in yourself and your organization. Boston, MA: Harvard Business School Publishing.

Shea, A. (2008, August 9). Reading the OED. On Point with Tom Ashbrook. [Audio Podcast] Retrieved from

Stanovich, K. E. (2009). What intelligence tests miss: The psychology of rational thought. New Haven, CT: Yale University Press.

Weick, K. E. (1979). The social psychology of organizing (2nd ed.). Reading, MA: Addison-Wesley.

Ryan Smerek is an assistant professor and assistant director of the MS in Learning and Organizational Change program at Northwestern University. He is the author of Organizational Learning and Performance: The Science and Practice of Building a Learning Culture.

Photo Source: Kelly Sikkema/Unsplash


“Oh no, not another ‘learning experience’”

I saw the bumper sticker once while driving in Boston some 15 years ago that said:

“Oh no, not another ‘learning experience.’”

I still smile at the sentiment. We know we should learn from failure but euphemisms such as having a “learning experience” seem tone deaf when faced with the many disappointments and setbacks we face. Sure, at some point we’ll make sense of the disappointment, but immediately glorifying failure as a “learning experience” doesn’t seem to respect the disappointment itself.

I never met the owner of the car with the bumper sticker, but you have to wonder what sentiment and life experience led them to joyfully paste that phrase on the back of his or her car. There is a forlorn stance with the opening “oh no,” and a humorous refusal of facing yet again another “learning experience.”

But the phrase indicates what we’ve all experienced, that many of the most critical things we have learned in life have come from failure. In fact, when I ask students to write a reflective paper on what they have learned from failure, it can be tough to even identify “pure” failures in our lives. What was once a failure becomes transformed as we reinterpret the past and take alternative actions that lead to redemptive paths. We “failed” by getting rejected by one school, but were accepted at another and flourished.

Failure is a momentary label, and its impact can be transformed by what is learned—both individually and in the organizations in which we work. So in the face of disappointments, mistakes, errors, and failures, see the developmental potential in them, and while you still might bemoan another “learning experience,” make it exactly that.

Ryan Smerek is an assistant professor and assistant director of the MS in Learning and Organizational Change program at Northwestern University. He is the author of Organizational Learning and Performance: The Science and Practice of Building a Learning Culture.

Photo Source: Sven Scheuermeier/Unsplash