JOHN WILCOX
  • Home
  • Curriculum Vitae
  • Teaching
    • Accuracy of Human Judgment
    • Introduction to Psychology
    • Teaching Methods
    • Philosophy of Science
    • Ethics in a Human Life
    • Epistemology & Probability
    • Logic
    • Applied Research Methods
    • Teaching Evaluations
  • John's Blog

John's Blog

How can we measure the accuracy of judgments and determine which ones to trust?

8/2/2024

0 Comments

 
(Forthcoming in Psychology Today)
​

​THE TL;DR KEY POINTS

  • I recently published an argument for calibrationism, the idea that judgments about the world are trustworthy only if—among other things—there’s evidence that they are produced in ways that are “well calibrated”
  • A set of judgments are well calibrated just in case they assign, say, 80% probabilities to things which are true 80% of the time, 90% probabilities to things which are true 90% of the time and so on
  • In this post, I provide a tool for measuring calibration, and I share some ideas about what to do and what not to do when measuring calibration, using my own experience as an example 

PRELIMINARY CLARIFICATION: WHAT IS CALIBRATION AND CALIBRATIONISM?
​
​As I mentioned elsewhere, I recently published a paper arguing for calibrationism, the idea that judgments of probability are trustworthy only if there’s evidence they are produced in ways that are calibrated—that is, only if there is evidence that the things one assigns probabilities of, say, 90% to happen approximately 90% of the time.

Below is an example of such evidence; it is a graph which depicts the calibration of a forecaster from the Good Judgment project—user 3559:
Picture
​The graph shows how often the things they assign probabilities to turn out to be true. For example, the top right dot represents all the unique events which they assigned a probability of around 97.5% to before they did or didn’t occur: that Mozambique would experience an onset of insurgency between October 2013 and March 2014, that France would deliver a Mistral-class ship to a particular country before January 1st, 2015 and so on for 17 other events. Now, out of all of these 19 events which they assigned a probability of about 97%, it turns out that about 95% of those events occurred. Likewise, if you look at all the events this person assigned a probability of approximately 0%, it turns out that about 0% of those events occurred. 

However, not all people are like this, below is a particular individual, user 4566, who assigned probabilities of around 97% to things which were true merely 21% of time, such as Chad experiencing insurgency by March 2014 and so on.

Read More
0 Comments

When should we trust the judgments from ourselves or others? The calibrationist answer

8/2/2024

0 Comments

 
(Forthcoming in Psychology Today)
​

​THE TL;DR KEY POINTS

  • We all make judgments of probability and use these to inform our decision-making
  • But it is not obvious which judgments to trust, and bad outcomes occur if we get it wrong—such as fatal misdiagnoses or false death sentence convictions
  • How do we determine which judgments to trust—both from ourselves or others?
  • I recently argued for inclusive calibrationism, which gives a two-part answer to this question
  • The first part says judgments of probability are trustworthy only if there’s evidence they are produced in ways that are "well calibrated"—that is, only if there is evidence that the things one assigns probabilities of, say, 90% to happen approximately 90% of the time
  • The second part says that judgments of probability are trustworthy only if they are also inclusive of all the relevant evidence
  • This blogpost then shares some ideas for implementing calibrationism, such as measuring calibration and creating evidence checklists to figure out how inclusive a judgment is

​THE IMPORTANCE OF TRUSTWORTHY JUDGMENTS

We all make judgments of probability and depend on them for our decision-making.

However, it is not always obvious which judgments to trust, especially since a range of studies suggest these judgments can sometimes be more inaccurate than we might hope or expect. For example, scholars have argued at least 4% of death sentence convictions in the US are false convictions, that tens or even hundreds of thousands of Americans die of misdiagnoses each year and that sometimes experts can be 100% sure of predictions which turn out to be false 19% of the time. So we want trustworthy judgments, or else bad outcomes can occur.

How do, then, can we determine which judgments to trust—either from ourselves or others? In a paper recently published here and freely available here, I argue for an answer called “inclusive calibrationism”—or just “calibrationism” for short. Calibrationism says trustworthiness requires two ingredients—calibration and inclusivity. 
​

Read More
0 Comments

    Author

    John Wilcox

    Cognitive scientist
    @ Columbia University
    Founder
    @ Alethic Innovations

    Archives

    March 2025
    December 2024
    September 2024
    August 2024
    January 2023
    November 2021
    May 2021
    July 2020

    Categories

    All
    Bayesianism
    Calibrationism
    Education
    Heuristics And Biases
    Judgment Accuracy
    Philosophy Of Science
    Politics And Governance
    Probability And Statistics
    Rationality
    Social Transformation

    RSS Feed

  • Home
  • Curriculum Vitae
  • Teaching
    • Accuracy of Human Judgment
    • Introduction to Psychology
    • Teaching Methods
    • Philosophy of Science
    • Ethics in a Human Life
    • Epistemology & Probability
    • Logic
    • Applied Research Methods
    • Teaching Evaluations
  • John's Blog