<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:wfw="http://wellformedweb.org/CommentAPI/" xmlns:dc="http://purl.org/dc/elements/1.1/" >

<channel><title><![CDATA[JOHN WILCOX - John\'s Blog]]></title><link><![CDATA[https://www.johnwilcox.org/johns-blog]]></link><description><![CDATA[John\'s Blog]]></description><pubDate>Sat, 04 Apr 2026 05:04:32 -0700</pubDate><generator>Weebly</generator><item><title><![CDATA[​Comparing the Constitutions of New Zealand and the United States]]></title><link><![CDATA[https://www.johnwilcox.org/johns-blog/comparing-the-constitutions-of-new-zealand-and-the-united-states]]></link><comments><![CDATA[https://www.johnwilcox.org/johns-blog/comparing-the-constitutions-of-new-zealand-and-the-united-states#comments]]></comments><pubDate>Sat, 22 Mar 2025 21:33:33 GMT</pubDate><category><![CDATA[Politics and governance]]></category><guid isPermaLink="false">https://www.johnwilcox.org/johns-blog/comparing-the-constitutions-of-new-zealand-and-the-united-states</guid><description><![CDATA[I have recently been studying the constitutions of the United States and New Zealand. I use&nbsp; the term &ldquo;constitution&rdquo; here to mean the foundational principles upon which their governments are built.&nbsp;Each country has quite different constitutions. The United States has its constitution built on the law entitled&mdash;quite aptly&mdash;the &ldquo;Constitution of the United States&rdquo;, arguably alongside the &ldquo;Declaration of Independence&rdquo;. In contrast, New Zealand [...] ]]></description><content:encoded><![CDATA[<div class="paragraph"><span style="color:rgb(123, 123, 123)">I have recently been studying the constitutions of the United States and New Zealand. I use&nbsp; the term &ldquo;constitution&rdquo; here to mean the foundational principles upon which their governments are built.&nbsp;</span><br /><br /><span style="color:rgb(123, 123, 123)">Each country has quite different constitutions. The United States has its constitution built on the law entitled&mdash;quite aptly&mdash;the &ldquo;Constitution of the United States&rdquo;, arguably alongside the &ldquo;Declaration of Independence&rdquo;. In contrast, New Zealand&rsquo;s constitution is comprised of various statutes, constitutional conventions and other important constitutional sources&mdash;salient among which are the Constitution Act 1986, the New Zealand Bill of Rights Act 1990 and the Treaty of Waitangi.</span><br /><br /><span style="color:rgb(123, 123, 123)">I want to highlight several themes which strike me, starting with the United States&rsquo; constitution.</span></div>  <div>  <!--BLOG_SUMMARY_END--></div>  <div class="paragraph" style="text-align:justify;"><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp;</em><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp;</em><span style="color:rgb(123, 123, 123)">&#8203;</span><em style="color:rgb(123, 123, 123)">&nbsp; U</em><em>S focus on welfare and happiness</em><br /><br />Central to the US constitution are concerns for welfare and happiness. This can be seen in several places, as italicized below:<br /><br /><ul><li>The Constitution states that &ldquo;We the People of the United States&hellip; do ordain and establish this Constitution for the United States of America&rdquo; in order to &ldquo;promote the general <em>Welfare</em>&rdquo; (among other ends such as &ldquo;Tranquility&rdquo;, &ldquo;Justice&rdquo; and &ldquo;common defense&rdquo;).</li><li>Again, the Constitution states that Congress has the power to collect taxes in order to &ldquo;to pay the Debts and provide for the common Defence and general <em>Welfare </em>of the United States&rdquo;.</li><li>Likewise, the Declaration of Independence states that &ldquo;Governments are instituted among Men&rdquo; to secure &ldquo;certain unalienable Rights&rdquo; including &ldquo;Life, Liberty and the pursuit of <em>Happiness</em>&rdquo;.</li><li>It further declares that &ldquo;whenever any Form of Government becomes destructive of these ends&rdquo;, then the people have the right to &ldquo;alter or abolish&rdquo; that government in favor of a new government predicated on the principles and powers that &ldquo;to them shall seem most likely to effect their Safety and <em>Happiness</em>&rdquo;.</li></ul><br />Interestingly, then, references to welfare and happiness are central to the founding principles of the United States<span style="color:rgb(123, 123, 123)">--</span>but not so much New Zealand.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp;</em><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp;</em><span style="color:rgb(123, 123, 123)">&#8203;</span><em style="color:rgb(123, 123, 123)">&nbsp;&nbsp;</em><em>Common focus on human rights</em><br /><br />Nevertheless, both countries have an emphasis on rights. For example, the Bill of Rights Act 1990 enshrines numerous rights, such as freedom of speech (Section 14) and thought and freedom from torture (Section 9) or unjust deprivation of life (Section 8). Other components of New Zealand&rsquo;s constitution&mdash;such as the Constitution Act 1986&mdash;focus less on the ethical dimension of government and instead specify the structure of its administrative machinery.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp;</em><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp;</em><span style="color:rgb(123, 123, 123)">&#8203;</span><em style="color:rgb(123, 123, 123)">&nbsp;&nbsp;</em><em>Common focus on public service</em><br /><br />Perhaps needless to say, both constitutions explicitly state that the purposes of government are not self-serving. Instead, the government exists to serve the people, such as by securing human rights or happiness. The government does not exist to serve the interests of the governors at the expense of the governed, so to speak.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp;</em><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp;</em><span style="color:rgb(123, 123, 123)">&#8203;</span><em style="color:rgb(123, 123, 123)">&nbsp;&nbsp;</em><em>NZ focus on indigenous partnership</em><br /><br />One way in which New Zealand differs, however, is that the Treaty of Waitangi&mdash;or its principles&mdash;is often regarded as a foundational component of New Zealand&rsquo;s government and constitution. At least in principle, the Treaty reflects an explicit partnership between, and regard for, the indigenous people of New Zealand&mdash;the Maori people. Interestingly, according to the Treaty of Waitangi Act 1975, &ldquo;Maori&rdquo; is statutorily defined as &ldquo;a person of the Maori race of New Zealand&rdquo;, including &ldquo;any descendant of such a person&rdquo;. The implication is that if one has <em>any</em> Maori blood&mdash;no matter how little&mdash;they are legally defined as Maori under that act.&nbsp;<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp;</em><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp;</em><span style="color:rgb(123, 123, 123)">&#8203;</span><em style="color:rgb(123, 123, 123)">&nbsp;&nbsp;</em><em>US focus on grievances</em><br /><br />Unlike New Zealand, however, the constitution of the United States has an explicit emphasis on grievances&mdash;27 in fact&mdash;which, in the eyes of the Founding Fathers, justified the creation of the United States. Understandably, these grievances involved violations of the very purpose government that is endorsed by the US constitution, such as refusal to assent to laws which the colonies deemed necessary &ldquo;for the public good&rdquo;, or the unpunished murder of two Maryland residents by British marines. That said, a lack of representation in the UK&rsquo;s parliament is only one of these grievances; for that reason, it is not entirely accurate to say American independence was motivated solely by the issue of &ldquo;taxation without representation&rdquo;&mdash;although that issue was undoubtedly an important one.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp;</em><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp;</em><span style="color:rgb(123, 123, 123)">&#8203;</span><em style="color:rgb(123, 123, 123)">&nbsp;&nbsp;</em><em>US focus on God</em><br /><br />Lastly, unlike New Zealand, the US constitution has an explicit emphasis on God. As one example, the Declaration concludes with a statement of reliance of God, especially since signing it was considered an act of treason that was punishable by death:<br />&#8203;</div>  <div><div class="wsite-multicol"><div class="wsite-multicol-table-wrap" style="margin:0 -15px;"> 	<table class="wsite-multicol-table"> 		<tbody class="wsite-multicol-tbody"> 			<tr class="wsite-multicol-tr"> 				<td class="wsite-multicol-col" style="width:8.786231884058%; padding:0 15px;"> 					 						  <div class="wsite-spacer" style="height:50px;"></div>   					 				</td>				<td class="wsite-multicol-col" style="width:82.513796194302%; padding:0 15px;"> 					 						  <div class="paragraph"><em><span style="color:rgb(123, 123, 123)">&ldquo;And for the support of this Declaration, with a firm reliance on the protection of divine Providence, we mutually pledge to each other our Lives, our Fortunes and our sacred Honor.&rdquo;</span></em></div>   					 				</td>				<td class="wsite-multicol-col" style="width:8.6999719216398%; padding:0 15px;"> 					 						  <div class="wsite-spacer" style="height:50px;"></div>   					 				</td>			</tr> 		</tbody> 	</table> </div></div></div>]]></content:encoded></item><item><title><![CDATA[Killingsworth Re-analysis script]]></title><link><![CDATA[https://www.johnwilcox.org/johns-blog/killingsworth-re-analysis-script]]></link><comments><![CDATA[https://www.johnwilcox.org/johns-blog/killingsworth-re-analysis-script#comments]]></comments><pubDate>Sat, 15 Mar 2025 00:44:13 GMT</pubDate><category><![CDATA[Uncategorized]]></category><guid isPermaLink="false">https://www.johnwilcox.org/johns-blog/killingsworth-re-analysis-script</guid><description><![CDATA[I've just posted a piece&nbsp;in Psychology Today about money and happiness. For interested researchers,&nbsp;here is the data used in the analysis, and here is the script used to do the analysis itself. [...] ]]></description><content:encoded><![CDATA[<div class="paragraph">I've just posted <a href="https://www.psychologytoday.com/us/blog/rationality-judgment-and-decision-making/202503/does-money-really-make-us-happier" target="_blank">a piece</a>&nbsp;in <em>Psychology Today </em>about money and happiness. For interested researchers,&nbsp;<a href="https://osf.io/qye4a/" target="_blank">here</a> is the data used in the analysis, and <a href="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/analysis_script.rmd">here</a> is the script used to do the analysis itself.</div>]]></content:encoded></item><item><title><![CDATA[What is a “social impact scholar”?]]></title><link><![CDATA[https://www.johnwilcox.org/johns-blog/what-is-a-social-impact-scholar]]></link><comments><![CDATA[https://www.johnwilcox.org/johns-blog/what-is-a-social-impact-scholar#comments]]></comments><pubDate>Tue, 03 Dec 2024 02:19:29 GMT</pubDate><category><![CDATA[Uncategorized]]></category><guid isPermaLink="false">https://www.johnwilcox.org/johns-blog/what-is-a-social-impact-scholar</guid><description><![CDATA[(Reposted from the LSE Impact Blog)Having a positive impact beyond academia can often be seen as a requirement, rather than as a personal orientation to research and its potential to create social change.&nbsp;John E. Wilcox&nbsp;and&nbsp;Brandon Reynante&nbsp;reflect on their experience as social impact scholars and what it means for their research.Our adult lives have been largely devoted to research, so it makes sense to ask, &ldquo;Why do we do it?&rdquo;. Our response to this question may h [...] ]]></description><content:encoded><![CDATA[<div class="paragraph" style="text-align:left;"><em>(Reposted from the <a href="https://blogs.lse.ac.uk/impactofsocialsciences/2024/11/21/what-is-a-social-impact-scholar/" target="_blank">LSE Impact Blog</a>)<br /><br />Having a positive impact beyond academia can often be seen as a requirement, rather than as a personal orientation to research and its potential to create social change.&nbsp;</em><span>John E. Wilcox&nbsp;</span><em>and&nbsp;</em><span>Brandon Reynante&nbsp;</span><em>reflect on their experience as social impact scholars and what it means for their research.</em><br /><br />Our adult lives have been largely devoted to research, so it makes sense to ask, &ldquo;Why do we do it?&rdquo;. Our response to this question may have changed over time, but our current answer is &ldquo;social impact&rdquo;&mdash;that is, to have a positive impact on society (assuming some understanding of that term).<br /><br />As we use the term, then, a &ldquo;social impact scholar&rdquo; is anyone who undertakes research largely with this objective of social impact in mind.&nbsp;<span>For example, we currently do research on how education can better prepare our societies for severe climate change in the future.</span>&nbsp;<span>We also know many others admirably working on social impact through, for instance, research that aims to improve the justice system or adolescent mental health.</span><br /></div>  <div>  <!--BLOG_SUMMARY_END--></div>  <div class="paragraph"><br /><span style="color:rgb(123, 123, 123)">A social impact scholar differs from what we might call a &ldquo;traditional scholar&rdquo; who may pursue research for a range of other reasons, such as pure intellectual curiosity, filling gaps in the literature or solving puzzles, regardless of whether doing so has any social impact in the broader sense.</span><br /><br /><br /><strong style="color:rgb(123, 123, 123)">Why is social impact scholarship important?</strong><br /><br /><span style="color:rgb(123, 123, 123)">While traditional scholarship is valuable and has its place, there are also various reasons why one might be a social impact scholar. The most obvious reason is that it can benefit society: other things being equal, being a successful social impact scholar by definition results in more good for others, and so it is good to pursue it for that reason. However, another reason is more self-centred: a&nbsp;</span><a href="https://doi.org/10.1016/j.jesp.2018.02.014" target="_blank">range of literature</a><span style="color:rgb(123, 123, 123)">&nbsp;suggests humans can derive happiness and fulfilment from helping and improving the lives of others, so social impact scholarship potentially benefits the scholars themselves as well. Then there are also institutional reasons social impact might matter for scholars: for example, many scholars may need to demonstrate social impact to secure funding or academic positions, as is the case with the UK&rsquo;s Research Excellence Framework.</span><br /><br /><br /><strong style="color:rgb(123, 123, 123)">How is social impact scholarship different?</strong><br /><br /><span style="color:rgb(123, 123, 123)">Defined as such, however, how does being a social impact scholar differ more specifically from traditional scholarship? Below is a non-exhaustive list of differences.</span><br /><br /><br /><span style="color:rgb(123, 123, 123)"><em>Difference #1: Success metrics</em></span><br /><br /><span style="color:rgb(123, 123, 123)">For a start, the success metrics of social impact scholarship differ. For traditional scholars, they might base the success of their endeavours on a range of traditional criteria: how many publications or citations they have, or the prestige of their places of publication or affiliation, for example. For social impact scholars, in contrast, the ultimate success criterion is societal good, and if anything else matters, then it is only in virtue of its (eventual) contribution to this goal. They seek to maximise their&nbsp;</span><em style="color:rgb(123, 123, 123)">social impact factor&nbsp;</em><span style="color:rgb(123, 123, 123)">rather than any traditional impact factor, like an&nbsp;</span><em style="color:rgb(123, 123, 123)">h-index</em><span style="color:rgb(123, 123, 123)">. Although such impact might not be straightforwardly measurable, obviously it could include diverse outcomes ranging from lives saved by reducing medical misdiagnoses to emissions offset by developing climate-mitigation technologies, to use some examples in our own fields. As it turns out, then, maximising social impact has a range of implications for other aspects of such scholarship, as follows.</span><br /><br /><br /><span style="color:rgb(123, 123, 123)"><em>Difference #2: Research agenda formation</em></span><br /><br /><span style="color:rgb(123, 123, 123)">For instance, one&rsquo;s research agenda often forms in a specific way: namely, by identifying some societal good one is aiming for, such as reducing fatal misdiagnoses or delivering more accurate and fair criminal convictions, and then working backwards from that goal to determine what research is necessary to achieve it. This problem orientation can then give rise to other differences from traditional scholarship.</span><br /><br /><br /><span style="color:rgb(123, 123, 123)"><em>Difference #3: Accessibility</em></span><br /><br /><span style="color:rgb(123, 123, 123)">In order for one&rsquo;s research to have a positive impact on some aspect of society, it needs to be accessible to those who can deliver that impact. This can have two more specific implications.</span><br /><span style="color:rgb(123, 123, 123)">First, sometimes the research must be&nbsp;</span><em style="color:rgb(123, 123, 123)">accessible as in comprehensible:</em><span style="color:rgb(123, 123, 123)">&nbsp;it should often be written in a way that non-specialists can understand what the research is saying and how it is important, for instance. This can in turn require non-technical language, clear examples or detailed explanations which considerably lengthen the work. For instance, some of our work aims to carefully explain concepts, like&nbsp;</span><a href="https://www.psychologytoday.com/intl/blog/rationality-judgment-and-decision-making/202410/when-should-we-trust-our-own-and-others" target="_blank">calibration</a><span style="color:rgb(123, 123, 123)">, which are familiar to specialists but not non-specialists.</span><br /><br /><span style="color:rgb(123, 123, 123)">Second, it needs to be&nbsp;</span><em style="color:rgb(123, 123, 123)">accessible as in available</em><span style="color:rgb(123, 123, 123)">&nbsp;to the relevant audiences. These requirements can mean that social impact scholarship requires one&nbsp;</span><a href="https://blogs.lse.ac.uk/impactofsocialsciences/2021/10/11/less-prestigious-journals-can-contain-more-diverse-research-by-citing-them-we-can-shape-a-more-just-politics-of-citation/" target="_blank">to publish in outlets that are not the most prestigious</a><span style="color:rgb(123, 123, 123)">&nbsp;if, for instance, more prestigious outlets are behind inaccessible paywalls or require concision in ways which alienate non-specialists.</span><br /><br /><br /><span style="color:rgb(123, 123, 123)"><em>Difference #4: Collaboration</em></span><br /><br /><span style="color:rgb(123, 123, 123)">Another difference of social impact scholarship is that it often benefits from&nbsp;</span><em style="color:rgb(123, 123, 123)">collaboration</em><span style="color:rgb(123, 123, 123)">&nbsp;with others, both so that others can contribute to the quality of the research and also so that others will be more likely to implement it for social good in their own contexts. Such collaboration may not only be with academics; it may be with public servants, volunteers or other impact-makers in the relevant domains. In this way, social impact scholarship is similar to community-engaged research and transdisciplinary research</span><em style="color:rgb(123, 123, 123)">.</em><span style="color:rgb(123, 123, 123)">&nbsp;In some areas, such as psychology, such collaboration is common, but sometimes it conflicts with aims of traditional scholarship. In traditional scholarship, for instance, if one wants to get recognition in their field for their work, then single-authorship is the best indication of their own unique contributions.</span><br /><br /><br /><strong style="color:rgb(123, 123, 123)">How is social impact scholarship&nbsp;<em>not&nbsp;</em>different?</strong><br /><br /><span style="color:rgb(123, 123, 123)">That said, there are also arguably a number of ways which in which social impact scholars should&nbsp;</span><em style="color:rgb(123, 123, 123)">not&nbsp;</em><span style="color:rgb(123, 123, 123)">be different to traditional scholars. For a start, they should have no less rigour than other scholars; after all, rigour is often necessary to form accurate judgments about what really drives positive societal impact. Additionally, social impact scholars shouldn&rsquo;t necessarily esteem themselves as being better people than traditional scholars: regardless of the accuracy of such an estimation, arrogance is arguably counter-productive, because if the aim of social impact scholarship is working with others to achieve social good, then a big ego may only alienate others.</span><br /><br /><span style="color:rgb(123, 123, 123)">Of course, some might accuse social impact scholars of merely virtue signalling or pursuing impact for impure reasons, such as to enhance one&rsquo;s own reputation. However, such accusations can be welcomed as tests of one&rsquo;s purity if one still pursues social impact&nbsp;</span><em style="color:rgb(123, 123, 123)">even though it invites such accusations which ironically might undermine their reputation</em><span style="color:rgb(123, 123, 123)">. After all, what matters is the impact, not one&rsquo;s reputation or what others think of it.</span><br /><span style="color:rgb(123, 123, 123)">&#8203;</span><br /><span style="color:rgb(123, 123, 123)">That, then, is our take on social impact scholarship, one that we hope might help others fruitfully reflect on whether or how they might like to make an impact with their research.</span></div>]]></content:encoded></item><item><title><![CDATA[Why being a JDM scholar is sometimes hard—really hard]]></title><link><![CDATA[https://www.johnwilcox.org/johns-blog/why-being-a-jdm-scholar-is-sometimes-hard-really-hard]]></link><comments><![CDATA[https://www.johnwilcox.org/johns-blog/why-being-a-jdm-scholar-is-sometimes-hard-really-hard#comments]]></comments><pubDate>Sat, 07 Sep 2024 06:26:32 GMT</pubDate><category><![CDATA[Uncategorized]]></category><guid isPermaLink="false">https://www.johnwilcox.org/johns-blog/why-being-a-jdm-scholar-is-sometimes-hard-really-hard</guid><description><![CDATA[TL;DR KEY POINTS  Judgment and decision making scholars&mdash;or "JDM scholars", for short&mdash;study how to improve judgment and decision making in diverse domainsDespite their motivation or ability to help others in these domains, life can sometimes be tough for them, and for three reasons:People sometimes think JDM scholars are&nbsp;useless because they are generalists who others might allege lack "specialization" or "experience" in their own&nbsp;specific domainPeople sometimes&nbsp;think J [...] ]]></description><content:encoded><![CDATA[<h2 class="wsite-content-title"><span style="color:rgb(0, 0, 0)">TL;DR KEY POINTS</span></h2>  <div class="paragraph"><ul><li><strong><em>Judgment and decision making scholars</em></strong><strong style="color:rgb(123, 123, 123)"></strong><span style="color:rgb(123, 123, 123)">&mdash;</span><strong style="color:rgb(123, 123, 123)"></strong><strong><em>or "JDM scholars", for short</em></strong><strong style="color:rgb(123, 123, 123)"></strong><span style="color:rgb(123, 123, 123)">&mdash;</span><strong style="color:rgb(123, 123, 123)"></strong><strong><em>study how to improve judgment and decision making in diverse domains</em></strong></li><li><strong><em>Despite their motivation or ability to help others in these domains, life can sometimes be tough for them, and for three reasons:</em></strong><ol><li><strong><em>People sometimes think JDM scholars are&nbsp;useless because they are generalists who others might allege lack "specialization" or "experience" in their own&nbsp;specific domain</em></strong></li><li><strong><em>People sometimes&nbsp;think JDM scholars are wrong&nbsp;because those people are psychologically programmed to think in intuitively compelling but demonstrably unreliable ways which JDM scholars are trained to avoid (even though JDM are&nbsp;also sometimes wrong too)</em></strong></li><li><strong><em>People sometimes hate JDM scholars because the JDM scholars think those people are wrong, and sometimes in ways with negative consequences</em></strong></li></ol></li><li><strong><em>Nevertheless, JDM scholars might make life easier for themselves in a couple of ways:</em></strong><ol><li><strong><em>&#8203;By finding so-called "JDM champions" who have the time, patience and open-minded humility to advocate JDM science in their domain</em></strong></li><li><strong><em>By aiming to make their communications accessible and tactful<span style="color:rgb(123, 123, 123)">&mdash;specifically by avoiding criticism and by fostering&nbsp;positivity</span></em></strong>&#8203;</li></ol></li></ul></div>  <div><div style="height: 20px; overflow: hidden; width: 100%;"></div> <hr class="styled-hr" style="width:100%;"></hr> <div style="height: 20px; overflow: hidden; width: 100%;"></div></div>  <div class="paragraph"><font color="#2a2a2a"><strong><font size="5">THE BACKGROUND</font></strong><br /></font><br /></div>  <div class="paragraph">&#8203;My life is great in a lot of ways: I&rsquo;m lucky to have a wonderful family and many projects and activities which I enjoy, for example. Thankfully,&nbsp; then, I'm very happy with my life overall.&nbsp;<br /><br />But despite that, one thing about my life can at times be very difficult and frustrating: being a judgment and decision making scholar.&nbsp;<br /><br />A judgment and decision making scholar&mdash;or a &ldquo;JDM scholar&rdquo; for short&mdash;is someone who professionally studies judgment and decision making: that is, someone who studies how we do, or how we should, make judgments and decisions.&nbsp; I&rsquo;m a JDM scholar because I want to improve judgments and decisions, both the ones from myself and those from others in their domains.<br /><br />And there are many domains where judgment and decision making could be improved: a JDM scholar could reduce <a href="https://www.pnas.org/doi/abs/10.1073/pnas.1306417111" target="_blank">false death sentence convictions</a> in law, <a href="https://jamanetwork.com/journals/jama/article-abstract/1845204#google_vignette" target="_blank">fatal misdiagnoses</a> in medicine or disastrous policies in politics, to take a few of countless examples.&nbsp;<br /><br />But despite both the possibility and promise of improving judgment and decision making in these domains, there are many reasons why this can be difficult or impossible for JDM scholars.<br /><br />I will explore some of them here, as well as why I think they sometimes stem from assumptions that are specious&mdash;that is, superficially plausible but actually wrong.<br /><br />(A caveat, though: while this blogpost talks of "people", it is not written with any specific "people" in mind<span style="color:rgb(123, 123, 123)">--</span>unless otherwise stated.)<br />&#8203;</div>  <div>  <!--BLOG_SUMMARY_END--></div>  <div class="paragraph"><font color="#2a2a2a">&#8203;<br /><strong><font size="5">REASON #1: PEOPLE MAY THINK YOU ARE USELESS</font></strong></font><br /><br /></div>  <div class="paragraph">&#8203;JDM scholars are often generalists: they study how to improve judgment and decision making in ways that are "domain general"&mdash;that is, applicable across many domains rather than just one.<br /><br />However, judgment and decision making, as it happens in practice, always happens within a specific domain, whether it is medicine, law, politics or something else.&nbsp; There is simply no such thing as &ldquo;applied judgment and decision making&rdquo; without it &ldquo;applying&rdquo; to "some" specific domain.&nbsp;<br /><br />But because of this, any time a JDM scholar aims to improve JDM in a given domain, those within the domain could criticize the scholar as lacking any specialization in that domain.&nbsp; They might say, &ldquo;She&rsquo;s not a doctor&rdquo;, or &ldquo;a lawyer&rdquo;, or &ldquo;a political analyst&rdquo;, for example.&nbsp;<br /><br />Concomitant with this is often the claim that the JDM scholar lacks &ldquo;experience&rdquo; in the relevant domain, where experience is typically operationalized in terms of the number of years working in the given domain.&nbsp;<br /><br />This, then, is one reason why being a JDM scholar is difficult. Sometimes this reason is right; it's known meteorologists are often accurate about their domain, and it's not obvious to me that JDM scholars generally have much to contribute there.<br /><br />But sometimes this reason is specious.&nbsp;<br /><br />First, the available evidence suggests years of experience is a weak predictor of judgment accuracy in some domains; for example, <a href="https://press.princeton.edu/books/hardcover/9780691178288/expert-political-judgment" target="_blank">Philip Tetlock</a> found that years of experience in geopolitical forecasting didn&rsquo;t predict accuracy at all, while <a href="file:///C:/Users/john-/Downloads/Calibration%20Training%20-%20MOK_DRM%202023.pdf" target="_blank">another study</a> found it had only a mild influence. The problem with some (but not all) domains is that<span style="color:rgb(123, 123, 123)">&mdash;as I argue <a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">elsewhere</a>&mdash;</span>they do not rigorously track judgment or decision quality, nor what correlates with it. It is analogous to how pre-modern physicians practiced bloodletting for millennia without rigorously studying whether it worked. Only with the development of modern medical science did we realize what truly worked and what didn&rsquo;t.<br /><br />The same point holds with judgment and decision making, and that brings me to a second point: the same literature shows that other things aside from domain experience or specialization matter more. For example, the <a href="https://link.springer.com/chapter/10.1007/978-3-031-30085-1_6" target="_blank">literature suggests</a> far stronger predictors of accuracy in domains like geopolitics are past records of accuracy, <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3779404" target="_blank">use of reference classes</a> and other <a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">domain general features</a> that are studied by JDM scholars.&nbsp;<br /><br />So although the JDM scholar often lacks the specialization or personal experience for a specific domain, what they do have is insight into the decades of rigorous scientific research about how to best make judgments and decisions, and sometimes in the specific domains too.<br /><br />Consequently, while it is sometimes correct to claim that a JDM scholar is useless simply because they are unspecialized or inexperienced, as with meteorology, there are other cases where this claim is like a medieval bloodletting doctor claiming that a modern medical student is useless because the doctor has years of experience and specialization in practices whose efficacy has not been rigorously studied.<br /><br />Unfortunately, however, it gets worse than that for another reason&hellip;<br />&#8203;</div>  <div class="paragraph">&#8203;<br /><strong><font color="#2a2a2a" size="5">REASON #2: PEOPLE MAY THINK YOU ARE WRONG</font></strong><br /><br /></div>  <div class="paragraph">&#8203;Not only will people think you are useless, but people will think you are wrong.&nbsp;<br /><br />Of course, sometimes the reason for this is that JDM scholars are wrong, and as JDM scholars, we have to be genuinely open to letting others correct our views through open-minded debate. However, a lot of the time people think you are wrong for another reason: namely, that humans are programmed to think in ways which are intuitively compelling but demonstrably unreliable.&nbsp;<br /><br />This comes up in <a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">countless ways</a>&mdash;in availability biases, in representativeness biases and in many others.&nbsp; In my own work, it comes up especially strongly for the so-called <a href="https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias" target="_blank">Monty Hall problem</a>&mdash;a brain-teaser where my experiments found that <em>every </em>untrained participant got the wrong answer. And when I taught logic at Stanford, I clearly remember one of my own students insisting<span style="color:rgb(123, 123, 123)">&mdash;</span><em>so strongly</em>&mdash;on why I was wrong about the correct answer to the very problem which I specialized in!<br /><br />She, however, was one of my better critics: although she thought I was wrong, she was at least willing to engage and argue with me about why that is so.<br /><br />Much of the time, people simply assume you&rsquo;re wrong from a distance, foreclose any possibility of questioning or challenging that assumption, and then they move on.<br /><br />This is problematic, since sometimes seeing the correct answer or reasoning in a given context requires the time, patience and open-mindedness to debate or undertake some training&mdash;and even then it might not work. For example, only a third of the trained participants in my experiments got the right answers to the Monty Hall problem (and some of them probably cheated even then!).<br /><br />Note that this differs from other specializations. For cryptologists or engineers, for instance, it&rsquo;s obvious to the non-specialist that they do not know what the right answer is; the average person would have no idea how to decode a given string of symbols or how to engineer a rocket ship. It&rsquo;s different in JDM research because the non-specialist is psychologically programmed to think in ways where they believe they already know what the right judgment or decision is; as <a href="https://osf.io/preprints/psyarxiv/atjve/download" target="_blank">David Mandel puts it</a>, they already have an intuitive theory of how to make judgments and decisions.<br /><br />This is not to say those theories are never right; sometimes they are and can contribute beneficial insights, but there are also many ways in which they can potentially be wrong as well. Unfortunately, though, all the ways in which these theories could be wrong are not obvious: if they were, then JDM scholars wouldn&rsquo;t need to study them for years or decades at a time.&nbsp;<br /><br />But unfortunately, it gets even worse than that for another reason&hellip;<br />&#8203;</div>  <div class="paragraph">&#8203;<br /><strong><font color="#2a2a2a" size="5">REASON #3: PEOPLE HATE MAY HATE BECAUSE YOU THINK THEY'RE WRONG</font></strong><br /><br /></div>  <div class="paragraph">&#8203;Not only will some people think you are useless and wrong, but they will sometimes hate you because you think they are wrong.&nbsp; For example, if you are a good JDM scholar, then you <em>have</em> to think uninitiated people are wrong when they first encounter the Monty Hall problem, no matter how confident, competent or qualified they themselves think they are. And of course, many of them would dislike you for thinking they are wrong&mdash;both because it insults them and because you appear arrogant. They might also be convinced your qualifications contribute only arrogance to your outlook instead of merit to your views.<br /><br />And unfortunately, this same phenomenon can come up in many other domains, because the patterns of thinking that lead to incorrect or correct reasoning in the Monty Hall and other problems are also potentially present in countless other domains, including medicine, law and many more (as I discuss <a href="https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias" target="_blank">here</a>).<br /><br />Yet not only would they hate you for thinking they are wrong, but they might feel that if you are right about their wrongness, then their job security, self-perceived performance or other things might be at risk&mdash;in which case they could hate you even more. What&rsquo;s more, we could see how the hate would be only worse if you think there may be horrible consequences because of their wrongness&mdash;like how innocent people die because of false diagnoses or false death sentence convictions.<br /><br /></div>  <div class="paragraph"><br /><strong><font size="5" color="#2a2a2a">SUMMARY AND SOLUTIONS?</font></strong>&nbsp;<br /><br /></div>  <div class="paragraph">To summarize, then, being a JDM scholar is hard for various reasons: people may think you&rsquo;re useless because you&rsquo;re &ldquo;unspecialized&rdquo; or &ldquo;inexperienced&rdquo; in their domain, because they are psychologically programmed to think you are wrong and because you think they are wrong in ways that can make them hate you.<br />&nbsp;<br />Consequently, being a JDM scholar might entail that if you try to help others by explaining how their judgments and decisions could be improved in their domain, then they might just think you&rsquo;re a "useless, inexperienced, incorrect and arrogant [insert your preferred personal swearword here]"&mdash;far from an outcome that benefits anyone.<br />&nbsp;<br />That, then, brings me to the final part of this blog: what are the solutions?<br /><br />So far, my answer is mainly, &ldquo;Yeah, I don&rsquo;t know&rdquo;.<br /><br />That said, I surmise two things may help when trying to improve judgment and decision making in a given domain.<br />&nbsp;<br />The first is finding people in that domain with three things: 1) the time, 2) the patience and 3) the open-minded humility to explore how JDM science can benefit them. We might call such people the &ldquo;champions&rdquo; of those domains, like how Brian Nosek uses the term to refer to the advocates of &ldquo;open science&rdquo; in specific disciplines. (In fact, that more I think about it, the more I realize all my JDM complaints might apply to metascience scholars too.) Unfortunately, however, many people lack at least one of the qualities to be champions of JDM science in their domain, and sometimes through no fault of their own&mdash;often we are too busy to take on extra work, for example. Whether there are JDM champions in a given domain is not something JDM scholars can easily control, however.<br />&nbsp;<br />The second thing that JDM scholars can do<span style="color:rgb(123, 123, 123)">--</span>and something that they <em>can</em> control<span style="color:rgb(123, 123, 123)">--</span>is how to make their communications as tactful and accessible as possible.<br />&nbsp;<br />Typically such tact may often require passing two tests. The first is the "no criticism" test: for example, no one should read your work and feel criticized by it. Sometimes such criticism is appropriate&mdash;like in a typical philosophy department&mdash;but sometimes it only counter-productively alienates people. The second test is the "positivity" test: people should read your work and feel positive afterwards.<br /><br />And additionally, JDM work should often pass an "accessibility" test: a non-specialist should be able to read your work and understand both what you are saying and why it might add value to their domain.<br />&nbsp;<br />However, these things are not easy, and I myself am guilty of violating both of these requirements at times.<br />&nbsp;<br />But as JDM scholars, perhaps aspiring to these standards might make life easier for us, as well as improving our likelihood of benefitting those we aspire to help.</div>]]></content:encoded></item><item><title><![CDATA[The seven requirements of highly accurate Bayesians]]></title><link><![CDATA[https://www.johnwilcox.org/johns-blog/the-seven-requirements-of-highly-accurate-bayesians]]></link><comments><![CDATA[https://www.johnwilcox.org/johns-blog/the-seven-requirements-of-highly-accurate-bayesians#comments]]></comments><pubDate>Fri, 09 Aug 2024 07:32:45 GMT</pubDate><category><![CDATA[Bayesianism]]></category><category><![CDATA[Philosophy of Science]]></category><category><![CDATA[Probability and statistics]]></category><guid isPermaLink="false">https://www.johnwilcox.org/johns-blog/the-seven-requirements-of-highly-accurate-bayesians</guid><description><![CDATA[TL;DR key points  We all form judgments about the world, and we need these to be accurate in order to make good decisionsIn epistemology and philosophy of science, &ldquo;Bayesianism&rdquo; is the dominating theory about how to form rational judgmentsHowever, not all of us are always Bayesians, and not all Bayesians are always accurate&nbsp;This post then articulates seven requirements to be an accurate Bayesian, some of which are widely-known to some (i.e. requirements 1 to 3) while others may  [...] ]]></description><content:encoded><![CDATA[<h2 class="wsite-content-title"><span><span style="color:rgb(0, 0, 0)">TL;DR key points</span></span></h2>  <div class="paragraph"><ul><li><em><strong>We all form judgments about the world, and we need these to be accurate in order to make good decisions</strong></em></li><li><em><strong>In epistemology and philosophy of science, &ldquo;Bayesianism&rdquo; is the dominating theory about how to form rational judgments</strong></em></li><li><em><strong>However, not all of us are always Bayesians, and not all Bayesians are always accurate&nbsp;</strong></em></li><li><em><strong>This post then articulates seven requirements to be an accurate Bayesian, some of which are widely-known to some (i.e. requirements 1 to 3) while others may be less so (i.e. requirements 4 to 7)</strong></em></li><li><em><strong>The requirements are as follows:</strong></em></li></ul><em><strong>&nbsp;</strong></em><em style="color:rgb(123, 123, 123)"><strong>&nbsp;&nbsp; &nbsp;&nbsp;</strong></em><em><strong> &nbsp; &nbsp; &nbsp; &nbsp; 1. Assign likelihoods to evidence</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp;</strong></em><em style="color:rgb(123, 123, 123)"><strong>&nbsp;&nbsp; &nbsp;&nbsp;</strong></em><em style="color:rgb(123, 123, 123)"><strong> &nbsp; &nbsp; &nbsp;&nbsp;</strong></em><em><strong>2. Assign prior probabilities to the hypotheses</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; </strong></em><em style="color:rgb(123, 123, 123)"><strong>&nbsp;&nbsp; &nbsp;&nbsp;</strong></em><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp;&nbsp;</strong></em><em><strong>3. Update using Bayes&rsquo; theorem</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; </strong></em><em style="color:rgb(123, 123, 123)"><strong>&nbsp;&nbsp; &nbsp;&nbsp;</strong></em><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp;&nbsp;</strong></em><em><strong>4. Use calibrated probabilities</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp;</strong></em><em style="color:rgb(123, 123, 123)"><strong>&nbsp;&nbsp; &nbsp;&nbsp;</strong></em><em style="color:rgb(123, 123, 123)"><strong> &nbsp; &nbsp;&nbsp;</strong></em><em><strong>5. Recognize auxiliary hypotheses</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp;</strong></em><em style="color:rgb(123, 123, 123)"><strong>&nbsp;&nbsp; &nbsp;&nbsp;</strong></em><em style="color:rgb(123, 123, 123)"><strong> &nbsp; &nbsp;&nbsp;</strong></em><em><strong>6. Recognize consilience</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; </strong></em><em style="color:rgb(123, 123, 123)"><strong>&nbsp;&nbsp; &nbsp;&nbsp;</strong></em><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp;&nbsp;</strong></em><em><strong>7. Be cautious about fallible heuristics<br /></strong></em></div>  <div><div style="height: 20px; overflow: hidden; width: 100%;"></div> <hr class="styled-hr" style="width:100%;"></hr> <div style="height: 20px; overflow: hidden; width: 100%;"></div></div>  <div class="paragraph">&#8203;<strong style="color:rgb(123, 123, 123)"><font color="#2a2a2a" size="5">THE IMPORTANCE OF BAYESIANISM&#8203;</font></strong></div>  <div class="paragraph"><br />As discussed elsewhere, in many important contexts, we need to form accurate judgments about the world: this is true of medical diagnosis and treatment, of law proceedings, of policy analysis and indeed of a myriad other domains. And as <a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">discussed elsewhere</a>, more accurate judgments often means better decisions, including in contexts where they can be a matter of life and death&mdash;such as medicine and law.<br /><br />In analytic epistemology and philosophy of science, &ldquo;Bayesianism&rdquo; is the dominant theory of how we should form rational judgments of probability. Additionally, as <a href="https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias" target="_blank">I discuss elsewhere</a>, Bayesian thinking can help us recognize strong evidence and find the truth in cases where others cannot.&nbsp;<br /><br />But there&rsquo;s ample evidence that humans are not Bayesians, and there&rsquo;s ample arguments that Bayesians can still end up with inaccurate judgments if they start from the wrong place (i.e. the wrong &ldquo;priors&rdquo;).<br /><br />So, given the importance of accurate judgments and given Bayesianism&rsquo;s potential to facilitate such accuracy, how can one be an accurate Bayesian?<br /><br />Here, I argue that there are seven requirements of highly accurate Bayesians (somewhat carrying on the Steven Covey-styled characterization of rationality which I outlined <a href="https://www.johnwilcox.org/johns-blog/the-seven-irrational-habits-of-highly-rational-people" target="_blank">here</a>). Some requirements will be well-known to relevant experts (such as requirements 1 to 3) while others might be less so (such as requirements 4 to 7). In any case, this post is written for both the expert and novice, hoping to say something unfamiliar to both&mdash;while the familiar remainder can be easily skipped.<br /><br />With that caveat, let us consider the first requirement.<br /><br /></div>  <div>  <!--BLOG_SUMMARY_END--></div>  <div class="paragraph" style="text-align:left;"><font size="5"><span style="color:rgb(123, 123, 123)">&#8203;</span><strong style="color:rgb(123, 123, 123)"><font color="#2a2a2a">THE SEVEN REQUIREMENTS OF HIGHLY ACCURATE BAYESIANS&#8203;</font></strong></font></div>  <div class="paragraph"><br /><em>&nbsp; &nbsp;</em><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp;</em>&#8203;<em> &nbsp;Requirement #1: Assigning Likelihoods&nbsp;</em><br /><br />Bayesianism requires us to consider the likelihood of the evidence we have given various hypotheses. Here, a &ldquo;likelihood&rdquo; is a technical term that refers to the probability of the evidence assuming the truth of some hypothesis that we are uncertain about. Examples include the likelihood of an improvement in your symptoms given the uncertain hypothesis that your medication is effective, the likelihood someone would smile at you given the uncertain hypothesis that they have a crush on you, or the likelihood someone&rsquo;s fingerprints would be at the scene of the crime given the uncertain hypothesis that they are guilty.<br /><br />In another <a href="https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias" target="_blank">blogpost</a>, I discuss evidence humans sometimes don&rsquo;t do this: they &ldquo;neglect&rdquo; the likelihoods in ways which make them oblivious of even strong evidence when they see it. For example, in one experiment, participants are aware that the evidence is 10 times more likely if one hypothesis is true compared to another, but when they receive that evidence, they assign the hypotheses equal probabilities instead assigning the correct probability of 91% to one of them. There, I also discuss how this evidence about likelihoods can arise in more realistic and natural settings too.&nbsp;<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp;&nbsp; &nbsp; &nbsp; &nbsp;</em><em>Requirement #2: Assigning Priors&nbsp;</em><br /><br />The second requirement is to assign prior probabilities to hypotheses, especially since sometimes multiple implausible hypotheses can make the evidence likely. For example, suppose you are at home, seemingly alone, and you hear the door slam. This could be quite likely if there is an intruder in your house, in which case it should somewhat raise your probability that there is such an intruder in the house.&nbsp;<br /><br />However, if there is a sufficiently low prior probability of an intruder, and if other explanations have a higher prior probability&mdash;like the wind slamming the door shut&mdash;then the evidence won&rsquo;t necessarily make the intruder hypotheses probable all things considered, even though it somewhat raises its probability.<br />&#8203;<br />This illustrates that when we consider the likelihood of the evidence, we also need to consider the prior probability of the relevant explanations in order to know just how confident we can be in those explanations&mdash;as I do in the three "realistic" examples <a href="https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias" target="_blank">here</a>.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp;&nbsp; &nbsp; &nbsp; &nbsp;</em><em>Requirement #3: Updating Using Bayes&rsquo; Theorem</em><br /><br />But Bayesianism tells us not just to consider likelihoods and to assign priors, but also to update our probabilities in specific ways once we receive our evidence. This often requires us to use Bayes&rsquo; theorem, since&nbsp;<span style="color:rgb(123, 123, 123)">if the priors and likelihoods are accurate, then <em>the only accurate posterior probability is the one prescribed by Bayes' theorem</em></span><em>.</em>&nbsp;Updating with Bayes' theorem can be complicated, but I give some examples of how this can be done <a href="https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias" target="_blank">here</a>, as well as a calculator <a href="https://www.johnwilcox.org/johns-blog/how-to-calculate-probabilities-the-bayesian-calculator" target="_blank">here</a> to easily facilitate these calculations.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;</em><em>Requirement #4: Using Calibrated Probabilities</em><br /><br />So Bayesianism requires us to assign initial probabilities&mdash;including prior probabilities for hypotheses and likelihoods for evidence&mdash;and to then update our probabilities when we receive that evidence.<br /><br />However, we also need to assign initial probabilities that are in some sense &ldquo;correct&rdquo;&mdash;an issue relating to the notorious &ldquo;problem of the priors&rdquo;. After all, if we assign likelihoods and prior probabilities, and these turn out to be inaccurate, then the outputs of the Bayes&rsquo; theorem will also be inaccurate.&nbsp;<br /><br /><a href="https://www.johnwilcox.org/johns-blog/when-should-we-trust-judgments-from-ourselves-or-others-the-calibrationist-answer" target="_blank">Elsewhere</a>, I argue for a particular solution to this problem called <em>calibrationism</em>. Calibrationism basically says we should trust probabilities<span style="color:rgb(123, 123, 123)">&mdash;</span>including the input probabilities in Bayesian calculations<span style="color:rgb(123, 123, 123)">&mdash;</span>only if we have evidence that the ways we have assigned them are "well calibrated".&nbsp;&nbsp;<br /><br />Probabilities are well calibrated when they correspond to the objective frequency with which things are true; for example, <a href="https://www.johnwilcox.org/johns-blog/how-can-we-measure-the-accuracy-of-judgments-and-determine-which-ones-to-trust" target="_blank">here</a> I discuss examples of forecasters who assign 97% probabilities to events which happen about 95% of the time&mdash;even events that appear &ldquo;unique&rdquo; and that might have looked like we couldn&rsquo;t assign numerically precise probabilities to. It is in principle easy to get this evidence about how calibrated we are; I explain how to do so in this other post <a href="https://www.johnwilcox.org/johns-blog/how-can-we-measure-the-accuracy-of-judgments-and-determine-which-ones-to-trust" target="_blank">here</a>.<br /><br />The implication is that if we have evidence that we assign calibrated probabilities, and we input these probabilities into Bayesian calculations, then we can likewise trust the outputs of these calculations, even in cases where the calculations instruct us to be highly confident or certain in sometimes counter-intuitive ways (like with the new Monty Hall problem <a href="https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias" target="_blank">here</a>).<br /><br />We can also verify the trustworthiness of outputs from Bayesian calculations in the same calibrationist way too: that is, by seeing whether the resulting probabilities similarly correspond to the frequency with which things are true. (And we can mathematically prove that they will be under certain assumptions.)<br /><br />There are various well-supported ways to get more accurate and calibrated probabilities, as I discuss <a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">here</a>. One is to use statistics or frequencies where relevant: for example, in assigning prior probabilities about whether the wind or an intruder caused the door to slam, we could consider statistics about how often you or others have experienced intruders, how often the wind slams doors shut in your house and so on. (Of course, we need not meticulously record these statistics, but even our experience can roughly tell us something about what the statistics would be, such as wind slamming the door every few days when the windows are open.)<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp;&nbsp; &nbsp; &nbsp;</em><em>Requirement #5: Recognizing Auxiliary Hypotheses and Successful Accommodation</em><br /><br />But being an accurate Bayesian also requires us to understand so-called &ldquo;auxiliary hypotheses&rdquo; and the role they play in trying to accommodate evidence. Auxiliary hypotheses are hypotheses that are distinct from the central ones we care about but which nevertheless may have implications for how we interpret the evidence.<br /><br />For example, <a href="https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias" target="_blank">here </a>I use an example where there is a low likelihood of a person sitting at the table in front of you if they did not have a crush on you.<br /><br />To make this likelihood appear higher, however, you might appeal to the auxiliary hypothesis that, say, that particular table just happens to be their favorite table. Now, we could suppose that this likelihood is higher: if the person does not have a crush on you <em>and</em> that table is their favorite table, then there is a 100% likelihood they would sit at that table. In this case, one might argue the evidence isn&rsquo;t so strong when they can appeal to an auxiliary hypothesis like this.<br /><br />In <a href="https://philarchive.org/archive/WILAHA-8?fbclid=IwAR0XcQqblntWdTUdYILhmu_nbduQ4bfewlnYFVWEP-ryd44h_vZBaVh6fb4" target="_blank">another article</a>, I discuss in detail how to make sense of attempts to accommodate evidence with auxiliary hypotheses like this. However, the main message is that an attempt like this would fail if the initial probability of the auxiliary hypothesis is sufficiently low and it will succeed if the initial probability of the auxiliary hypothesis is sufficiently high.<br /><br />For example, if we have no independent evidence or good reason to think that that particular table is their favorite table out of the 64 other tables, then we can assign it a 1/65 initial probability. And if that is the case, then this accommodation attempt fails because it attempts to raise the likelihood of the evidence only by appealing to an improbable auxiliary hypothesis.<br /><br />We can prove this using a theorem from that paper called the <em>theorem of successful accommodation</em>. The theorem basically says that an auxiliary hypothesis (symbolized with <em>a</em>) and central hypothesis (symbolized with <em>h<font size="1">1</font></em>) accommodate some evidence (symbolized with <em>e</em>) as successfully as&mdash;or more successfully than&mdash;some alternative hypothesis <em>h<font size="2">2</font></em> just in case:</div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/ad-hocness_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div class="paragraph">&#8203;In words, this says that the accommodation attempt is successful just in case the likelihood of the evidence given <em style="color:rgb(123, 123, 123)">h<font size="2">2</font></em> is less than or equal to the result of multiplying the initial probability of the auxiliary hypothesis (supposing <em style="color:rgb(123, 123, 123)">h<font size="1">1</font></em> is true) and the likelihood of the evidence given <em style="color:rgb(123, 123, 123)">h<font size="1">1</font></em> and the auxiliary <em>a</em>.&nbsp;<br /><br />In this case, we could suppose the evidence<em> e</em> is that the person sits at the table in front of you, <em>P(e|<em style="color:rgb(123, 123, 123)">h<font size="2">2</font></em>)</em> is the likelihood they would sit at that table (denoted with <em>e</em>) if they have a crush on you (denoted with <em style="color:rgb(123, 123, 123)">h<font size="2">2</font></em>), <em>P(e|<em style="color:rgb(123, 123, 123)">h<font size="1">1</font></em>&amp;a)</em> is the likelihood they would sit at that table if they did not have a crush on you (denoted with <em style="color:rgb(123, 123, 123)">h<font size="1">1</font></em>) but that table was still their favorite table (denoted with <em>a</em>) and <em>P(a|<em style="color:rgb(123, 123, 123)">h<font size="1">1</font></em>)</em> is the probability that that table is their favorite table (assuming they do not have a crush on you).&nbsp;<br /><br />And if we suppose that we have no reason to think that table is their favorite table, then <em>P(a|<em style="color:rgb(123, 123, 123)">h<font size="1">1</font></em>)</em> is low, we could plug in some probabilities as below and the evidence strongly favors the hypothesis that they have a crush on you:</div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/ad-hocness-example_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div class="paragraph">In this way, we could use these theorems and principles to determine when appealing to auxiliary hypotheses successfully accommodates the evidence in this and countless other cases.<br /><br />&#8203;<br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp;</em><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp;</em><em>Requirement #6: Recognizing Consilience</em><br /><br />Auxiliary hypotheses play an especially important role when a central hypothesis is potentially supported by multiple kinds of evidence&mdash;in which case we say the central hypothesis &ldquo;consiliates&rdquo; the evidence. I also discuss consilience in detail that <a href="https://philarchive.org/archive/WILAHA-8?fbclid=IwAR0XcQqblntWdTUdYILhmu_nbduQ4bfewlnYFVWEP-ryd44h_vZBaVh6fb4" target="_blank">article</a>.&nbsp;<br /><br />There, a main take away is that sometimes the evidence will strongly support a hypothesis which explains multiple kinds of evidence if the probability of the alternative hypotheses are sufficiently low&mdash;even if each alternative hypothesis seems potentially plausible when considered in isolation.<br /><br />I use one example to illustrate this: Darwin&rsquo;s arguments in favor of evolution over special creationism (the idea that God created each species in a separate creative act).<br /><br />Darwin thought evolution was probable because it could explain diverse evidence: <ul><li>on his theory, some so-called &ldquo;non-aquatic&rdquo; birds had webbed feet even though they didn&rsquo;t go in or near water because they evolved from other birds with webbed feet that were in the water; <br /></li><li>many animals (such as frogs and lizards) share similar bone structures because they evolved from common ancestors; <br /></li><li>blind insects inside European caves are more similar to those outside of their caves than to insects in similar caves elsewhere in the world, and this is because they evolved from the insects immediately outside their caves; </li></ul>...and so on and so forth.<br /><br />However, Darwin noted that the special creationist could in principle offer another explanation for each piece of evidence: that God simply desired it to be that way&mdash;that God simply desired the similarities in webbed feet among specific birds, the similarities in bone structures among some animals, and the similarities among some insects&mdash;even though he created them all separately.<br /><br />As I discuss in the article, the problem with the creationist explanation is that, for each piece of evidence, it appeals to an auxiliary assumption about God&rsquo;s desires that&mdash;while potentially plausible in isolation&mdash;is still less than certain. And combining uncertainty with uncertainty results in even more uncertainty.&nbsp;<br /><br />We can use an analogy to illustrate the point: if each auxiliary hypothesis has the probability of a coin flip landing heads&mdash;50%&mdash;then the probability they are all true is <em>much lower</em>, even though they each might seem plausible or possible in isolation. For example, the probability of three coin-flips landing heads is 50%&times;50%&times;50%=12.5%, and one can be 87.5% confident that not all coin flips landed heads.&nbsp;<br /><br />Appealing to independent multiple auxiliary hypotheses to accommodate evidence likewise faces the same problem. If those auxiliary hypotheses are not sufficiently supported, then the appeal will fail, regardless of whether it&rsquo;s auxiliaries about whether creationism is true, or about whether someone does not have a crush on you, or about something else&mdash;even when those auxiliary hypotheses might seem plausible in isolation. And when this is the case, then the evidence can favor a consiliating alternative explanation&mdash;like evolution&mdash;that explains multiple kinds of evidence.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp;</em><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp;</em><em>Requirement #7: Caution about Pervasive but Fallible Heuristics</em><br /><br />Above, I&rsquo;ve discussed many ideas about how to assign probabilities, but there&rsquo;s ample evidence that we often assign probabilities using so-called <em>heuristics</em>&mdash;that is, efficient but often sub-optimal thinking processes. And while sometimes useful, they can cause us to assign inaccurate probabilities.<br /><br />Below, then, are a set of heuristics to be cautious of:<br /><br /></div>  <div><div class="wsite-multicol"><div class="wsite-multicol-table-wrap" style="margin:0 -15px;"> 	<table class="wsite-multicol-table"> 		<tbody class="wsite-multicol-tbody"> 			<tr class="wsite-multicol-tr"> 				<td class="wsite-multicol-col" style="width:12.08519068846%; padding:0 15px;"> 					 						  <div class="wsite-spacer" style="height:50px;"></div>   					 				</td>				<td class="wsite-multicol-col" style="width:77.180234449699%; padding:0 15px;"> 					 						  <div class="paragraph"><br /><strong style="color:rgb(123, 123, 123)">Availability heuristic:</strong><span style="color:rgb(123, 123, 123)">&nbsp;People often assign probabilities or estimates on the basis of how mentally available something is&mdash;that is, how readily relevant instances come to mind. For example, in one study, participants thought more people died of floods than asthma, arguably because dying by floods is more mentally available than dying by asthma and arguably because floods are more prevalent in the media. However, the reverse was true and asthma actually claimed 9 times as many lives as floods.</span><br /><br /><strong style="color:rgb(123, 123, 123)">Representativeness</strong><span style="color:rgb(123, 123, 123)">&nbsp;<strong>heuristic:</strong> People often assign a probability to something based on how similar that thing seems to a typical instance of the relevant category. For example, multiple studies find many participants think a woman named Linda is more likely to be a feminist bank teller than a bank teller in general who may or may not be a feminist. But this is incorrect since Linda cannot be more likely to be in a smaller group of people (such as feminist bank tellers) than in a larger group which contains that smaller group (such as all bank tellers in general). To make it more concrete, imagine there are 30,000 feminist bank tellers and 60,000 bank tellers in general (which comprises the 30,000 feminist bank tellers plus the 30,000 non-feminist bank tellers). Clearly Linda cannot be more likely to be in the smaller group of 30,000 bank tellers than in the group of 60,000 bank tellers which includes that smaller group.</span><br /><br /><strong style="color:rgb(123, 123, 123)">Consistency or chance heuristic:</strong><span style="color:rgb(123, 123, 123)">&nbsp;While perhaps not as widely documented, my experiments suggest participants often evaluate the probability of competing hypotheses based on whether they are each consistent with the evidence: if, in principle, either hypothesis is potentially true given the evidence&mdash;and they are therefore consistent with the evidence in that respect&mdash;then assign them an equal probability, or so the thinking goes. Participants might further think that if something is possible given chance&mdash;or consistent with it&mdash;then it&rsquo;s just as probable in light of the evidence. This seems to be what people are doing in the <a href="https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias" target="_blank">Monty Hall experiments</a> where they assign the hypotheses equal but inaccurate probabilities: door C being opened is consistent with door A or door B concealing the prize, and Monty Hall could have just happened to open door C by chance, even if there was&nbsp; a 10% likelihood of him doing so given door A concealing the prize, so the hypotheses are equally. As we have seen, though, this thinking leads to the wrong answer.</span><br /><br /></div>   					 				</td>				<td class="wsite-multicol-col" style="width:10.734574861842%; padding:0 15px;"> 					 						  <div class="wsite-spacer" style="height:50px;"></div>   					 				</td>			</tr> 		</tbody> 	</table> </div></div></div>  <div class="paragraph"><span style="color:rgb(123, 123, 123)">Consequently, we should be careful in thinking something is probable or improbable merely because we can or can&rsquo;t easily call relevant instances to mind (as per availability) or because it resembles typical instances (as per representativeness). We also need to avoid thinking hypotheses are equally probable merely because they are consistent with the evidence.&nbsp;</span><br /><br /><span style="color:rgb(123, 123, 123)">Instead, if we are cautious about these heuristics and meet the other six requirements, then I think we will be highly accurate Bayesians.</span></div>]]></content:encoded></item><item><title><![CDATA[The seven “irrational” habits of highly rational people]]></title><link><![CDATA[https://www.johnwilcox.org/johns-blog/the-seven-irrational-habits-of-highly-rational-people]]></link><comments><![CDATA[https://www.johnwilcox.org/johns-blog/the-seven-irrational-habits-of-highly-rational-people#comments]]></comments><pubDate>Fri, 09 Aug 2024 05:28:58 GMT</pubDate><category><![CDATA[Heuristics and biases]]></category><category><![CDATA[Probability and statistics]]></category><category><![CDATA[Rationality]]></category><guid isPermaLink="false">https://www.johnwilcox.org/johns-blog/the-seven-irrational-habits-of-highly-rational-people</guid><description><![CDATA[(Forthcoming in&nbsp;Psychology Today)&#8203;  TL;DR key points  We all make judgments about the world and then use these to make important decisionsBut if someone made perfectly accurate judgments and sound decisions, would we recognize it?The science and philosophy of judgment and decision-making suggests that the answer is often &ldquo;no&rdquo;This post identifies 7 habits of highly rational people&mdash;habits which others often label as &ldquo;irrational&rdquo;:&nbsp; &nbsp; &nbsp; &nbsp;  [...] ]]></description><content:encoded><![CDATA[<div class="paragraph"><span style="color:rgb(123, 123, 123)">(Forthcoming in&nbsp;</span><em style="color:rgb(123, 123, 123)">Psychology Today</em><span style="color:rgb(123, 123, 123)">)<br />&#8203;</span><br /></div>  <h2 class="wsite-content-title"><span><span style="color:rgb(0, 0, 0)">TL;DR key points</span></span></h2>  <div class="paragraph"><ul><li><em><strong>We all make judgments about the world and then use these to make important decisions</strong></em></li><li><em><strong>But if someone made perfectly accurate judgments and sound decisions, would we recognize it?</strong></em></li><li><em><strong>The science and philosophy of judgment and decision-making suggests that the answer is often &ldquo;no&rdquo;</strong></em></li><li><em><strong>This post identifies 7 habits of highly rational people&mdash;habits which others often label as &ldquo;irrational&rdquo;:</strong></em></li></ul><em><strong>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;1. Highly rational people are confident in things despite &ldquo;no good evidence&rdquo; for them</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;</strong></em><em><strong>2. They are confident in things which are outright false</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;</strong></em><em><strong>3. They countenance the &ldquo;impossible&rdquo; and are &ldquo;paranoid&rdquo;</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;</strong></em><em><strong>4. They avoid risks that don&rsquo;t happen</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;</strong></em><em><strong>5. They pursue opportunities that fail</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;</strong></em><em><strong>6. They are often irrational&nbsp;</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;</strong></em><em><strong>7. They do things that are often &ldquo;crazy&rdquo; or &ldquo;unconventional&rdquo;&nbsp;</strong></em><ul><li><em><strong>I lastly conclude with some evidence-based suggestions about how we can distinguish the genuinely rational from the irrational:</strong></em></li></ul> <em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;</strong></em><em><strong>1. Measure calibration</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;</strong></em><em><strong>2. Learn norms of reasoning</strong></em><br /><em style="color:rgb(123, 123, 123)"><strong>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp; &nbsp;</strong></em><em><strong>3. Think in terms of expected utility theory</strong></em><ul><li><em><strong>This might help us to both recognize and make trustworthy judgments and decisions in our lives</strong></em></li></ul></div>  <div><div style="height: 20px; overflow: hidden; width: 100%;"></div> <hr class="styled-hr" style="width:100%;"></hr> <div style="height: 20px; overflow: hidden; width: 100%;"></div></div>  <div class="paragraph"><strong style="color:rgb(123, 123, 123)"><font color="#2a2a2a" size="5">THE IMPORTANCE OF RECOGNIZING WHAT'S RATIONAL AND WHAT'S NOT&#8203;</font></strong></div>  <div class="paragraph">&#8203;<br />If someone was as rational as could be&mdash;with many accurate and trustworthy judgments about the world, and with sound decisions&mdash;would we recognize it? There are reasons to think the answer is &ldquo;No&rdquo;. In this piece, I aim to challenge prevailing intuitions about rationality: I will argue that the philosophy and science of judgment and decision-making reveal a number of ways in which what <em>appears</em> to be rational diverges from what <em>actually is</em> rational.&nbsp;<br /><br />This piece takes its title from Steven Covey&rsquo;s well-known book &ldquo;The Seven Habits of Highly Effective People&rdquo;. I will argue that, similarly, there are seven habits of highly rational people&mdash;but these habits can appear so counter-intuitive that others label them as &ldquo;irrational&rdquo;. Of course, the rationality of these habits might be obvious to specialists in judgment and decision-making, but I find they are often not so obvious to others of the sort for whom this piece is written.<br /><br />In any case, not only are these habits potentially interesting in their own right, but recognizing them may also help to open our minds, to help us better understand the nature of rationality and to better identify the judgments and decisions we should trust&mdash;or not trust&mdash;in our own lives.<br /><br />Without further ado, then, I present&hellip;</div>  <div class="paragraph"><br /><font size="5">&#8203;<strong style="color:rgb(123, 123, 123)"><font color="#2a2a2a">THE SEVEN "IRRATIONAL" HABITS OF HIGHLY RATIONAL PEOPLE<br /></font></strong></font></div>  <div class="paragraph">&nbsp;<br />&#8203;<em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; 1. Highly rational people are confident in things despite &ldquo;no good evidence&rdquo; for them</em><br /><br /><span style="color:rgb(123, 123, 123)">The first habit of highly rational people is that they are sometimes confident in things when others think there is &ldquo;no good evidence&rdquo; for them.</span></div>  <div>  <!--BLOG_SUMMARY_END--></div>  <div class="paragraph"><br />One case where this shows up extremely clearly is the Monty Hall problem, as I discuss in detail in a blogpost <a href="https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias" target="_blank">here</a>.<br /><br />In the problem, a prize is randomly placed behind one of three doors, you select a door and then the gameshow host&mdash;named Monty Hall&mdash;will open one of the other doors that does not conceal the prize. If you select a door and that door conceals the prize, then Monty Hall will open either of the other two doors with an equal likelihood. But if the door you select does not conceal the prize, then Monty Hall must open the only other door that does not conceal the prize and which you did not select.<br /><br />In these circumstances, as I explain in the <a href="https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias" target="_blank">blogpost</a>, if you select door A and Monty Hall opens door C, then there&rsquo;s a 2/3 probability that door B conceals the prize. In this case, then, door C being opened constitutes &ldquo;evidence&rdquo; that door B conceals the prize. Furthermore, let us consider an adaptation called the &ldquo;new Monty Hall problem&rdquo;: in this case, door C would be opened with a 10% likelihood if door A conceals the prize, in which case there&rsquo;s provably a 91% probability that door B conceals the prize after door C is opened. In this version, the truly rational response is to be very confident that door B conceals the prize.<br /><br />But despite this, in my experiments, <em>everyone</em> without training who encountered these problems got the wrong answer, and the vast majority thought door B had only a 50% probability of concealing the prize in both versions of the problem. This effectively means they thought there was no good evidence for door B concealing the prize when there in fact was!<br /><br />What&rsquo;s more, <a href="https://www.cambridge.org/core/journals/judgment-and-decision-making/article/likelihood-neglect-bias-and-the-mental-simulations-approach-an-illustration-using-the-old-and-new-monty-hall-problems/B80822FC020A13C4A83E293C6120492E" target="_blank">the studies</a> found that not only did participants not recognize this good evidence, but they were also more confident in their incorrect answers. Compared to participants who were trained and more likely to get the correct answers to these problems, the other participants were on average both more confident in the correctness of their (actually incorrect) answers and they thought they had a better understanding of why their (actually incorrect) answers were correct.<br /><br />What this shows is that truly rational people may recognize objectively good evidence for hypotheses where others think there is none&mdash;leading them to be confident in things in ways which others think are irrational. In the <a href="https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias" target="_blank">blogpost</a>, I also discuss some more realistic scenarios where this in principle could occur<span style="color:rgb(123, 123, 123)">&mdash;</span>including some from medicine, law and daily life.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;</em><em>2. They are confident in things that are outright false</em><br /><br />But even if someone is rationally confident in something, that thing might be false a particular proportion of the time.<br /><br />In fact, according to one norm of trustworthy judgments, a perfectly accurate person would often be 90% confident in things that are false approximately 10% of the time. In other words, a perfectly accurate person would be &ldquo;well calibrated&rdquo; in the rough sense that, in normal circumstances, anything they assign 90% probabilities to will be true approximately 90% of the time, anything they assign 80% probabilities to will be true approximately 80% of the time and so on.<br /><br />We can see this when we look at well calibrated forecasters who might assign high probabilities to a bunch of unique events, and while most of those will happen, some of them will not&mdash;as I discuss in detail <a href="https://www.johnwilcox.org/johns-blog/how-can-we-measure-the-accuracy-of-judgments-and-determine-which-ones-to-trust" target="_blank">here</a>. Yet if we focus on a small sample of cases, they might look less rational than they are since they will be confident in things that are outright false.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;</em><em>3. They countenance the &ldquo;impossible&rdquo; and are &ldquo;paranoid&rdquo;</em><br /><br />However, <a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">studies</a> suggest many people&mdash;including experts with PhDs about their domain, doctors, jurors and the general public&mdash;are not so well calibrated. One example of this is <em>miscalibrated certainty</em>&mdash;that is, when people are certain (or virtually certain) of things which turn out to be false.<br /><br />For instance, <a href="https://press.princeton.edu/books/hardcover/9780691178288/expert-political-judgment" target="_blank">Philip Tetlock</a> tracked the accuracy of a group of political experts&rsquo; long-term predictions, and he found that out of all the things they were 100% certain would <u><strong>not</strong></u> occur, those things actually <u><strong>did</strong></u> occur 19% of the time. <a href="https://psycnet.apa.org/record/2011-15298-010" target="_blank">Other studies</a> likewise suggest people can be highly confident in things which are false a significant portion of the time.<br /><br />But a perfectly rational person wouldn&rsquo;t be so miscalibrated about these things which others are certain about, and so they would assign higher probabilities to things which others would think are &ldquo;impossible&rdquo;. For example, a perfectly calibrated person would perhaps assign 19% probabilities to the events which Tetlock&rsquo;s experts were inaccurately certain would not happen&mdash;or they might even assign some of them much higher probabilities, like 99%, if they had sufficiently good evidence for them. In such a case, the perfectly rational person would look quite &ldquo;irrational&rdquo; from the perspective of Tetlock&rsquo;s experts.<br /><br />But insofar as miscalibrated certainty is widespread among experts or the general public, so too would be the perception that truly rational people are &ldquo;irrational&rdquo; in virtue of them countenancing what others irrationally consider to be &ldquo;improbable&rdquo; at best or &ldquo;impossible&rdquo; at worst.<br /><br />Furthermore, when one has miscalibrated certainty about outcomes that are &ldquo;bad&rdquo;, not only will a rational person look like they believe in the possibility of &ldquo;impossible&rdquo; outcomes, but the rational person will look irrationally &ldquo;paranoid&rdquo; in doing so since the supposedly &ldquo;impossible&rdquo; outcomes are bad.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; &nbsp; &nbsp;</em><em style="color:rgb(123, 123, 123)">&nbsp;</em><em>4. They avoid risks that don&rsquo;t happen</em><br /><br />But not only will a rational person look &ldquo;irrational&rdquo; or &ldquo;paranoid&rdquo; in virtue of thinking the &ldquo;impossible&rdquo; is possible or even (probably) true, but they will also act to reduce risks which never actually happen.<br /><br />This is because our <a href="https://plato.stanford.edu/entries/rationality-normative-utility/" target="_blank">leading theory of rational decision-making</a> claims that we should make decisions not just based on how probable or improbable outcomes are, but rather based on the so-called <em>expected utility</em> of those outcomes, where the expected utility of an outcome is the product of both its probability and how good or bad it is.<br /><br />This sometimes entails that we should avoid decisions if there is an improbable chance of something really bad happening. For example, it can be rational to avoid playing Russian roulette even if the gun is unlikely to fire a bullet, simply because the off-chance of a bullet firing is <em>so bad</em> that this outcome has high expected negative utility.&nbsp;<br /><br />Likewise, for many other decisions in life, it may be rational to avoid decisions if they have improbable outcomes that are sufficiently bad. This consequently means a rational person could often act to avoid many risks that never actually happen.<br /><br />But as is well-known, people often evaluate the goodness of a decision based on its outcome, and if the bad thing does not happen, the average person might evaluate that decision as &ldquo;irrational&rdquo;.&nbsp;<br /><br />This kind of thing could happen quite often too: for example, if for every highly negative outcome with a probability of 10% and which a rational decision-maker would avoid, then they will be avoiding negative outcomes which simply don&rsquo;t happen 90% of the time, potentially making them look quite irrational.&nbsp;<br /><br />The situation is even worse if the evaluator has the "miscalibrated certainty" we considered earlier, and the outcome not only does not happen, but rather it looks like it was always &ldquo;impossible&rdquo; from the perspective of the evaluator.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;</em><em>5. They pursue opportunities that fail</em><br /><br />But the same moral holds not only for decisions that avoid risk, but also for decisions that pursue reward. For example, a rational decision-maker might accept an amazing job offer which has merely a 10% chance of continued employment if the possibility of continued employment is sufficiently good. But of course, that decision to accept that job has a 90% chance of resulting in unemployment, potentially making the decision again seem like a &ldquo;failure&rdquo; if the probable happens.<br /><br />More generally, a rational decision-maker would pursue risky options with 90% chances of failure if the options are sufficiently good all things considered: it is like buying a lottery ticket with a mere 10% chance of winning but with a sufficiently high reward.<br /><br />But again, the rational decision-maker could look highly &ldquo;irrational&rdquo; in the 90% of cases where those decisions lead to the less-than-ideal outcomes.&nbsp;<br /><br />In any case, what both this habit and the preceding one have in common is that rational decision-making requires making decisions that lead to the best outcomes <em>over many decisions in the long-run</em>, but humans often evaluate decision-making strategies based <em>on mere one-off cases</em>.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; </em><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp;&nbsp;</em><em>6. They are often irrational</em><br /><br />Despite that, though, arguably any realistic person who is as rational as could be would still be genuinely irrational to some degree. This is because our dominant theory of judgment and decision-making<span style="color:rgb(123, 123, 123)">&mdash;</span><em>dual process theory</em>&mdash;entails that while we often make reflective judgments and decisions, there are countless situations where we do not and simply cannot.&nbsp;<br /><br />Instead, the <a href="https://www.cambridge.org/core/books/heuristics-and-biases/0975337F864379F2729EAD873D804BA8" target="_blank">literature</a> commonly affirms that everyone employs a set of so-called <em>heuristics </em>for judgment and decision-making which&mdash;while often adequate&mdash;also often lead to sub-optimal outcomes. Consequently, even if someone was as rational as could be, they would still make irrational judgments and decisions in countless other contexts where they cannot be expected to rely on their more reflective faculties.<br /><br />If we then focus solely on these unreflective contexts, we would get an inaccurate impression of how rational they are overall.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp;&nbsp;</em><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp;&nbsp;</em><em>7. They do things that are often &ldquo;crazy&rdquo; or &ldquo;unconventional&rdquo;</em><br /><br />All of the preceding thoughts then entail that rational people may do things that seem &ldquo;crazy&rdquo; or &ldquo;unconventional&rdquo; by common standards: they might believe in seemingly impossible things, or act to reduce risks that never happen, or pursue opportunities that never materialize, and so on. This might express itself in weird habits, beliefs, or in many other ways.<br /><br />But this shouldn&rsquo;t be too surprising. After all, the history of humanity is a history of common practices which later generations appraise as unjustified or irrational. Large portions of humanity once believed that the earth was flat, that the earth was at the center of the universe, that women were supposedly incapable or unsuitable for voting, and so on and so forth.<br /><br />Have we then finally reached the apex of understanding in humanity&rsquo;s evolution, a point where everything we now do and say will appear perfectly rational by future standards? If history is anything to go by, then surely the answer is &ldquo;No&rdquo;. And if that is the case, then perhaps the truly "rational" will be ahead of the rest&mdash;believing or doing things which seem crazy or irrational by our currently common standards.<br />&#8203;</div>  <div class="paragraph"><span style="color:rgb(123, 123, 123)">&#8203;</span><br /><strong style="color:rgb(123, 123, 123)"><font color="#2a2a2a"><font size="5">HOW TO DISTINGUISH THE RATIONAL FROM THE IRRATIONAL</font></font></strong><br /><br /></div>  <div class="paragraph">What I hope to have conveyed, then, is just how frequently our untrained intuitions about what is rational may diverge from what is truly rational: what&rsquo;s rational might appear &ldquo;irrational&rdquo;, and vice versa. In a world where these intuitions might lead us astray, then, how can we really tell rational from irrational, accurate from inaccurate or wisdom from unwisdom?<br /><br />Some common rules of thumb might not work too well. For example, <a href="https://link.springer.com/chapter/10.1007/978-3-031-30085-1_6" target="_blank">sometimes the evidence</a> fails to find that years of experience, age or educational degrees improve accuracy, at least in domains like geopolitical forecasting.&nbsp;<br /><br />What follows, then, are some suggestions which I think are supported by the evidence:<br /><br /><br />&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; <em>Suggestion #1: Measure calibration</em><br /><br />First, track the calibration of the judgments you care about&mdash;whether they are yours or others. I provide some tools and ideas for how to do this <a href="https://www.johnwilcox.org/johns-blog/how-can-we-measure-the-accuracy-of-judgments-and-determine-which-ones-to-trust" target="_blank">here</a>. This can help us to put things in perspective, to avoid focusing on single-cases and to detect pervasive miscalibration which can afflict our decision-making. And as <a href="https://link.springer.com/chapter/10.1007/978-3-031-30085-1_6" target="_blank">other studies suggest</a>, past accuracy is the greatest predictor of future accuracy.<br /><br /><br /><span style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;</span><em>Suggestion #2: Learn norms of reasoning</em><br /><br />Additionally, I would suggest learning and practicing various norms of reasoning. These include the evidence-based suggestions for forming more accurate judgments in my book <em><a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">Human Judgment</a></em>, such as practicing active open-minded thinking and thinking in terms of statistics. It also includes other norms, such as so-called &ldquo;Bayesian reasoning&rdquo; which can produce more accurate judgments in the Monty Hall problem and potentially other contexts, as I discuss <a href="https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias" target="_blank">here</a>&nbsp;and <a href="https://www.johnwilcox.org/johns-blog/the-seven-requirements-of-highly-accurate-bayesians" target="_blank">here</a>.<br /><br /><br /><span style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;</span><em>Suggestion #3: Think in terms of expected utility</em><br /><br />Finally, when evaluating the rationality of someone&rsquo;s decisions, think in terms of expected utility theory. Expected utility theory is complicated, but <a href="https://plato.stanford.edu/entries/rationality-normative-utility/" target="_blank">here</a> is a potentially helpful introduction to it (and from my former PhD advisor<span style="color:rgb(123, 123, 123)">&mdash;a really awesome person!</span>). In short, though, expected utility theory requires us to ask what probabilities people attach to outcomes, how much they value those outcomes and&mdash;on my preferred version of it&mdash;whether their probabilities are calibrated and their values are in some sense objectively &ldquo;correct&rdquo;. Then, we can ask whether they are making decisions which lead to the best possible outcomes in the long-run.<br /><br />In these ways, I think we can better tell what&rsquo;s rational from what&rsquo;s not in a world where our intuitions can otherwise lead us astray.<br /><br /></div>]]></content:encoded></item><item><title><![CDATA[How to recognize good evidence and find the truth where others cannot: Identifying and overcoming “likelihood neglect bias”]]></title><link><![CDATA[https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias]]></link><comments><![CDATA[https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias#comments]]></comments><pubDate>Fri, 09 Aug 2024 04:59:01 GMT</pubDate><category><![CDATA[Bayesianism]]></category><category><![CDATA[Heuristics and biases]]></category><category><![CDATA[Probability and statistics]]></category><category><![CDATA[Rationality]]></category><guid isPermaLink="false">https://www.johnwilcox.org/johns-blog/how-to-recognize-good-evidence-and-find-the-truth-where-others-cannot-identifying-and-overcoming-likelihood-neglect-bias</guid><description><![CDATA[(Forthcoming in&nbsp;Psychology Today)&#8203;  &#8203;THE TL;DR KEY POINTS  As humans, biases often prevent us from forming accurate judgments about the worldIn a&nbsp;recently published set of experiments, I show how a newly introduced bias can also do this: likelihood neglect biasTo illustrate the bias and how to overcome it, I discuss how it arises in versions of the &ldquo;Monty Hall problem&rdquo;In some versions of the problem, the evidence can objectively favor a hypothesis with a probabi [...] ]]></description><content:encoded><![CDATA[<div class="paragraph"><span style="color:rgb(123, 123, 123)">(Forthcoming in&nbsp;</span><em style="color:rgb(123, 123, 123)">Psychology Today</em><span style="color:rgb(123, 123, 123)">)<br />&#8203;</span><br /></div>  <h2 class="wsite-content-title">&#8203;THE TL;DR KEY POINTS</h2>  <div class="paragraph"><ul><li><strong><em>As humans, biases often prevent us from forming accurate judgments about the world</em></strong></li><li><strong><em>In a&nbsp;<a href="https://www.cambridge.org/core/journals/judgment-and-decision-making/article/likelihood-neglect-bias-and-the-mental-simulations-approach-an-illustration-using-the-old-and-new-monty-hall-problems/B80822FC020A13C4A83E293C6120492E" target="_blank">recently published set of experiments</a>, I show how a newly introduced bias can also do this: likelihood neglect bias</em></strong></li><li><strong><em>To illustrate the bias and how to overcome it, I discuss how it arises in versions of the &ldquo;Monty Hall problem&rdquo;</em></strong></li><li><strong><em>In some versions of the problem, the evidence can objectively favor a hypothesis with a probability of 91%</em></strong></li><li><strong><em>However, the experiments show untrained participants are oblivious of this, they effectively think the evidence is irrelevant and so they do not recognize objectively strong evidence for the truth when presented with it</em></strong></li><li><strong><em>I then apply these ideas to other more realistic contexts&mdash;including medicine, law and mundane situations&mdash;to show how they can help us recognize good evidence and find the truth where others may not</em></strong></li></ul></div>  <div><div style="height: 20px; overflow: hidden; width: 100%;"></div> <hr class="styled-hr" style="width:100%;"></hr> <div style="height: 20px; overflow: hidden; width: 100%;"></div></div>  <div class="paragraph"><strong style="color:rgb(123, 123, 123)"><font color="#2a2a2a" size="5">THE IMPORTANCE OF RECOGNIZING GOOD EVIDENCE</font></strong></div>  <div class="paragraph"><br />&#8203;We all need to form accurate judgments about the world in many diverse and important contexts. What is the correct diagnosis for someone&rsquo;s medical condition? Does someone have a crush on you? Did the defendant kill the victim? Here, I will discuss how the evidence can reveal the truth about these questions<span style="color:rgb(123, 123, 123)">&mdash;</span>and potentially others which you might care about<span style="color:rgb(123, 123, 123)">&mdash;</span>but only if we think in the right ways.<br /><br />It&rsquo;s well-documented that various biases can hinder us in our quest for truth. In a recently published paper in <em>Judgment and Decision Making</em> (freely available <a href="https://www.cambridge.org/core/journals/judgment-and-decision-making/article/likelihood-neglect-bias-and-the-mental-simulations-approach-an-illustration-using-the-old-and-new-monty-hall-problems/B80822FC020A13C4A83E293C6120492E" target="_blank">here</a>), I introduce a new cognitive bias: likelihood neglect bias. Understanding this bias, and how to overcome it, can help us recognize good evidence and find the truth in numerous cases where others might not.<br /><br />To show this, though, I&rsquo;ll use a well-known brain-teaser which reveals this bias&mdash;the Monty Hall problem&mdash;and then I&rsquo;ll apply the emerging ideas to show how we can find the truth in other realistic cases&mdash;including medicine, law and more mundane topics. You might then want to apply these ideas to other cases which you might care about.</div>  <div>  <!--BLOG_SUMMARY_END--></div>  <div class="paragraph"><br /><strong style="color:rgb(123, 123, 123)"><font color="#2a2a2a" size="5">THE MONTY HALL PROBLEM<br />&#8203;</font></strong></div>  <div class="paragraph">&#8203;In the Monty Hall problem, a prize is randomly placed behind one of three doors. You select a door, which remains closed for the time being. If the prize is behind the door which you selected, then the gameshow host, Monty Hall, will randomly open one of the other two doors with an equal probability. But if the prize is behind one of the doors which you did not initially select, then he will open the other door which does not conceal the prize and which you did not select.<br /><br />In my experiments, participants read a description of the problem and then are asked to imagine that they select door A and that Monty Hall opens door C, like in the very high-quality picture below:</div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/mh_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div class="paragraph">&#8203;The question, then, is this: what is the probability that the prize is behind door B after Monty Hall opens door C?<br /><br />Most people think the answer is 1/2 or 50%. But actually, the correct answer is that it has a 2/3 or 66% probability of concealing the prize.<br /><br />However, virtually no one without training gets it right when they first encounter the problem, including my once younger self and all the 50 participants in one group of my experiments, for instance.&nbsp;<br /><br />In fact, it is because the right answer is so counter-intuitive that one cognitive scientist has called the Monty Hall problem &ldquo;the most expressive example of cognitive illusions or mental tunnels in which even the finest and best-trained minds get trapped&rdquo; (Piattelli-Palmarini, 1994a, p. 161).<br /><br />To see how the correct solution is indeed correct, though, let us consider the probabilistic explanation before applying the same ideas to some more important real-life problems.<br />&#8203;<br />&#8203;</div>  <div class="paragraph"><strong style="color:rgb(123, 123, 123)"><font color="#2a2a2a" size="5">LIKELIHOOD NEGLECT BIAS AND THE EXPLANATION OF THE MONTY HALL PROBLEM</font></strong></div>  <div class="paragraph"><br />&#8203;From a probabilistic standpoint, the sole reason door B more probably conceals the prize than door A is because of the likelihood that Monty Hall would open door C if door B conceals the prize. Here, a &ldquo;likelihood&rdquo; is a technical term used in probability theory: it refers to the probability of the evidence that we are certain about if we assume the truth of the hypothesis that we are not certain about. In this case, we are certain about door C being opened, but not about which door conceals the prize, so the likelihood refers to door C being opened given that a particular door conceals the prize.<br /><br />The likelihood of the evidence differs for whether door A or door B conceals the prize. If door A conceals the prize, then there&rsquo;s a 50% likelihood that Monty Hall would open door C, since he could have opened either door B or C. But if door B conceals the prize, then there&rsquo;s a 100% likelihood that Monty Hall would open door C, since he cannot open the door that you selected (door A) nor the door that conceals the prize (door B), so he must open the only remaining door (door C).<br /><br />From a probabilistic perspective, the fact that door C being opened is twice as likely if door B conceals the prize (and the fact the prize was randomly placed behind one of the doors) makes it the case that door B is twice as probable to conceal the prize than door A. Mathematicians calculate this using Bayes&rsquo; theorem as below (although you don&rsquo;t need to know Bayes&rsquo; theorem or anything aside from the importance of likelihoods here):<br /><br /></div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/yea-boi_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div class="paragraph" style="text-align:center;"><span style="color:rgb(123, 123, 123)">where<em> P(c)</em> is the probability door B conceals the prize given that door C is opened,&nbsp;</span><em style="color:rgb(123, 123, 123)">P(A), P(B)</em><span style="color:rgb(123, 123, 123)">&nbsp;and&nbsp;</span><em style="color:rgb(123, 123, 123)">P(C)</em><span style="color:rgb(123, 123, 123)">&nbsp;are all the prior probabilities that the respective doors conceal the prize and&nbsp;</span><em style="color:rgb(123, 123, 123)">P(c|A), P(c|B)</em><span style="color:rgb(123, 123, 123)">&nbsp; and&nbsp;</span><em style="color:rgb(123, 123, 123)">P(c|C)</em><span style="color:rgb(123, 123, 123)">&nbsp;are the likelihoods of door C being opened given the respective hypotheses.</span></div>  <div class="paragraph" style="text-align:left;"><br />Again, the formula isn&rsquo;t too important, but what is important is that door B more probably conceals the prize because of the likelihoods&mdash;because <em><span style="color:rgb(123, 123, 123)">P(c|A)</span></em>=50% and <em><span style="color:rgb(123, 123, 123)">P(c|B)</span></em>=100%.<br /><br />What my experiments showed, however, is that most participants were aware of what the likelihoods were, but they didn&rsquo;t realize that these likelihoods favored one hypothesis over another (and for other explanations of this and the Monty Hall problem, see the Appendix below). This, then, is likelihood neglect bias: by definition, it is being aware that the evidence is more likely given one hypothesis compared to another, all while failing to realize that the evidence therefore raises the probability of the former hypothesis relative to the latter.<br /><br />Technically, likelihood neglect bias is a violation of what is known as the <em>law of likelihood</em><span style="color:rgb(123, 123, 123)">&mdash;</span>a law of probability which states (roughly speaking) that if the evidence is more likely given one hypothesis compared to another, then the evidence necessarily raises the probability of the former hypothesis relative to the latter. The strength which the evidence provides for the hypothesis is measured by the so-called <em>likelihood ratio</em><span style="color:rgb(123, 123, 123)">&mdash;that is, how much more likely the evidence is assuming one hypothesis compared to the other</span>. In the original Monty Hall problem, the likelihood ratio is 100%/50%=2, meaning that door C is twice as likely to be opened if door B conceals the prize than if door A conceals the prize.&nbsp;<br /><br />Overcoming likelihood neglect bias requires us to apply the law of likelihood in our assessment of the evidence, thereby raising our probabilities for hypotheses and getting closer to the truth in cases where others do not.<br />&#8203;</div>  <div class="paragraph"><strong style="color:rgb(123, 123, 123)"><font color="#2a2a2a" size="5">RECOGNIZING STRONG EVIDENCE WHERE OTHERS DO NOT:<br />&#8203;LIKELIHOOD NEGLECT BIAS AND THE NEW MONTY HALL PROBLEM</font></strong></div>  <div class="paragraph"><br />&#8203;In a sense, then, overcoming likelihood neglect bias can potentially help us recognize strong evidence and find the (probable) truth where others can not.<br /><br />To see how this is, though, we could consider another innovation of the experiments: they showed that not only do people get incorrect probabilities for the original Monty Hall problem, but they also get the incorrect probabilities for an adaptation of the problem where there&rsquo;s objectively strong evidence to favor door B over door A.<br /><br />More specifically, in another experiment, participants were presented with what I call the &ldquo;new Monty Hall problem&rdquo;, a problem where the likelihood of door C being opened given that door A conceals the prize is merely 10%&mdash;not 50% like in the original Monty Hall problem. In this case, the likelihood ratio changes to 100%/10%=10, meaning door C is 10 times more likely to be opened if door B conceals the prize than if door A conceals the prize.<br /><br />Under these conditions, one can prove with mathematics and simulations that the probability that door B conceals the prize after door C is opened if 91%&mdash;as I show in <a href="https://www.cambridge.org/core/journals/judgment-and-decision-making/article/likelihood-neglect-bias-and-the-mental-simulations-approach-an-illustration-using-the-old-and-new-monty-hall-problems/B80822FC020A13C4A83E293C6120492E#article" target="_blank">the paper</a> and its <a href="https://static.cambridge.org/content/id/urn%3Acambridge.org%3Aid%3Aarticle%3AS1930297524000081/resource/name/S1930297524000081sup001.zip" target="_blank">supplementary materials</a>.<br /><br />But despite this, 84% of participants&mdash;or 21 of 25 of them&mdash;thought the evidence was effectively irrelevant and that door B had merely a 50% probability of concealing the prize. (And the four other participants gave flawed answers for different reasons too.)<br /><br />Importantly, though, most of the participants were aware of the likelihoods&mdash;they knew that door C being opened was much more likely given that door B conceals the prize than that door A conceals the prize&mdash;but remarkably, this did not affect their probabilities for the hypotheses at all.<br /><br />This, again, is likelihood neglect bias: it is realizing the evidence is more likely given one hypothesis than another, all while not realizing that this evidence favors that hypothesis over the other.<br /><br />What this shows, then, is that sometimes the world can present us with strong evidence to favor some hypotheses over others, yet we might be oblivious of how this is so if we neglect the likelihood of the evidence.<br /><br />The evidence could potentially favor hypotheses to the point of virtual certainty too, especially if, say, there are multiple pieces of evidence where we neglect the likelihoods. For instance, if there are just two pieces of evidence as strong as door C being opened in the new Monty Hall problem, then one can be 99% certain of the relevant hypothesis&mdash;and this is in a case where we know humans fail to think the evidence is relevant at all!<br /><br />The evidence could in principle be so strong as to virtually reveal the truth, but we might fail to realize this if we neglect the likelihoods and do not think about the evidence in the right ways.<br />&#8203;</div>  <div class="paragraph"><strong style="color:rgb(123, 123, 123)"><font color="#2a2a2a" size="5">EXAMPLES IN REALISTIC CONTEXTS&nbsp;</font></strong></div>  <div class="paragraph"><br />To see how this could be the case in more important and realistic scenarios, let us consider a few examples, each of which are adapted from real situations where subsequent evidence supported the favored hypothesis (albeit with the details changed to preserve anonymity).<br /><br /><br /><em>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; Example #1: Medical diagnosis and treatment</em><br /><br />Suppose Tabatha has a debilitating medical condition but does not know for sure what is causing it. Suppose she furthermore has some hunches and tries a particular therapy with some suggestive but severely limited studies supporting its efficacy. She then assigns the therapy an initial 10% probability of working, but she tries it for a few months, and she starts feeling much better.<br /><br />Now, if the therapy was effective, we can suppose there would be a 100% likelihood she would have a remarkable improvement (and I will discuss how we can assign these values later, although the exact values of the likelihoods do not matter so much as the general ideas).&nbsp;<br /><br />On the other hand, though, if the therapy was ineffective, how do we determine the likelihood she would have recovered by chance? One demonstrably accurate method of assigning probabilities is by looking at the frequency with which she has had similar improvements in her condition in the past, since study after study after study have suggested the past can help us assign probabilities.&nbsp;<br /><br />To do this, then, suppose Tabatha has had the disease for 3 years and has never seen a comparable improvement in her condition before. Then, using this historical information, she could accurately assign a likelihood of her remarkably improving given chance that is 1/156 or approximately 0.6%, since she had not improved remarkably for any of the past 156 months. In this case, we then have a likelihood ratio of 100%/0.6%=166, meaning the recovery is about 166 more times likely if the therapy is effective than if it is not.<br /><br />If this is the case, then using Bayes&rsquo; theorem (like one can in this post <a href="https://www.johnwilcox.org/johns-blog/how-to-calculate-probabilities-the-bayesian-calculator" target="_blank">here</a>), the probability that the therapy is effective given her remarkable recovery is approximately 95%, and she can be confident that the therapy works.&nbsp;<br /><br />However, if she commits likelihood neglect and instead attributes her improvement to mere &ldquo;chance&rdquo;, then she could fail to realize the implications of the evidence which strongly favors one hypothesis over another.&nbsp;<br /><br />Of course, one alternative explanation is that she is improving merely because of a placebo effect: perhaps her trying the treatment gives her the hope or belief that she will improve, and this itself causes the improvement. Placebo effects are real, and that is why there are widespread experimental protocols to rule them out.<br /><br />But in this case, we could suppose she has evidence to discredit this explanation. Perhaps, for example, she has had years of trying alternative therapies, some of which she was even more confident in because they were recommended by experts. But if those therapies did not improve her condition despite their potential for a placebo effect, then the probability of the placebo effect would be undermined by this evidence of historical frequencies.<br /><br />Likewise, when getting into details, there may also be evidence to undermine other alternatives, in which case Tabatha can be virtually certain the therapy is effective if she avoids likelihood neglect.<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;</em><em>Example #2: Someone having a crush on you</em><br /><br />Let us use another example which has&mdash;for educational purposes&mdash;turned out to engage my students&rsquo; interest in the past: whether someone has a crush on you.<br /><br />Suppose you and another person notice each other a lot in a dining hall at your college; you both make eye contact a lot and you assign a conservative 30% initial probability to them having a crush on you, since you have historically noticed that at least 30% of people who make eye contact with you like that have turned out to like you.<br /><br />Now suppose that one lunch time, the dining hall is virtually empty, there are 65 free tables around, but when this person comes in to get lunch, they choose to sit at the table directly in front of you.&nbsp;<br /><br />You are wondering, then, if them sitting in front of you means they like you, and again, the likelihoods could provide an answer.&nbsp; In particular, suppose that if they do not like you, then they could have sat at any of the 65 tables, and so the likelihood of them sitting at the table in front of you is 1/65 or 1.5%&mdash;very unlikely.<br /><br />However, if they like you, then we could suppose the likelihood of them sitting at the table in front of you is about 70%&mdash;and much more likely&mdash;because they might want to sit there to get your attention, to encourage you to approach them and start a conversation or whatever.&nbsp;<br /><br />If that is the case, then the likelihood ratio would be 70%/(1/65)=45.5 and would again mean that the evidence strongly favors the hypothesis that they like you&mdash;in this case with a probability of 95%. But again, if we merely attribute the occurrence to &ldquo;chance&rdquo; while neglecting the likelihoods, we might similarly not realize how strong the evidence is.<br /><br />&#8203;<br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;</em><em>Example #3: Law and the attribution of a crime</em><br /><br />Let us use another example that is based on an actual scenario (albeit with some details modified), this time concerning law. Jemaine is an investigative journalist based in central Africa, and he publishes a news piece that is critical of a local gang. Six months later, his house is burned down. However, it is known that some in the gang did not like the critical news piece, and there have been rumors they have retaliated against others before. What, then, is the probability that Jemaine&rsquo;s house was sabotaged?<br /><br />Again, if we rely on well-documented heuristics and assign the &ldquo;accident&rdquo; hypothesis a high probability merely because it seems plausible or consistent with the evidence, then we might commit likelihood neglect and fail to recognize evidence for the truth. To think more carefully about the case, we could assign initial probabilities and then consider likelihoods.<br /><br />There are multiple rumors that the gang has retaliated against critics before, and so suppose we assign an initial 10% probability to the hypothesis that Jemaine would experience retaliation before updating on the evidence about the house burning down. (That said, if the gang is sufficiently powerful and motivated, it could be very difficult to determine the true frequency with which they successfully execute covert arson.)&nbsp;<br /><br />Then, we can think about the likelihood of Jemaine&rsquo;s house burning down given that he was or wasn&rsquo;t retaliated against. Suppose that if he was retaliated against, there&rsquo;s a 20% likelihood he would have his house burned down, since there are also other ways in which he could have been retaliated against&mdash;such as a plane crash, staged suicide and so on. On the other hand, suppose that if he was not retaliated against, then there is a 1/10,000 likelihood his house would burn down, since although house burnings occasionally happen, they are still exceedingly improbable relative to the vastly larger number of cases where house burnings never occur. In this case, the likelihood ratio would be 20%/(1/10,000)=2,000, meaning the house burning down is 2,000 times more likely if he was retaliated against.<br /><br />Again, then, using Bayes&rsquo; theorem, we can be 99.5% confident Jemaine&rsquo;s house was burned down by the gang in central Africa. And in the real-life case which this example was adapted from, an informant close to the &ldquo;gang hierarchy&rdquo; did indeed confirm Jemaine was retaliated against despite the gang&rsquo;s attempts to cover it up. But again, if we neglect likelihoods and rely merely on what seems &ldquo;plausible&rdquo; or &ldquo;consistent&rdquo; with the evidence, we could fail to recognize strong evidence when it can reveal the truth to us.<br /><br /></div>  <div class="paragraph"><font size="5"><strong style="color: rgb(123, 123, 123);"><font color="#2a2a2a" style="">CONCLUDING THOUGHTS</font></strong>&#8203;</font></div>  <div class="paragraph"><br />&#8203;Of course, in the examples above, some people might naturally think the evidence is relevant and not neglect the likelihoods, and it&rsquo;s an open-question as to exactly when people neglect likelihoods. However, from experience, I know that some people can neglect likelihoods in cases like these, even if others do not.<br /><br />In any case, the point is that the experimental evidence suggests we can sometimes neglect likelihoods, potentially in more important contexts too.<br /><br />That said, an important warning: even if we do not neglect likelihoods, there are many ways we might try to reason about them unsuccessfully. For example, we might assign incorrect likelihoods in the first place, or we might not properly take into account so-called &ldquo;auxiliary hypotheses&rdquo;, or we might not use valid tools like Bayes&rsquo; theorem to correctly revise our views given information about likelihoods. I discuss how to avoid these and other errors in another <a href="https://www.johnwilcox.org/johns-blog/the-seven-requirements-of-highly-accurate-bayesians" target="_blank">post here</a>.&nbsp;<br /><br />For now, though, here&rsquo;s an appendix of other explanations for the Monty Hall problem.<br />&#8203;<br /><br /></div>  <div class="paragraph"><strong style="color:rgb(123, 123, 123)"><font color="#2a2a2a" size="5">APPENDIX: MORE EXPLANATIONS OF THE MONTY HALL PROBLEM</font></strong></div>  <div class="paragraph"><br />&#8203;Above, I gave an explanation of the Monty Hall problem in terms of probability theory, but there&rsquo;s other ways to explain the correct answer to the problem too.&nbsp;<br /><br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;</em><em>Explanation #2: The Many Doors Adaptation</em><br /><br />One style of explanation comes from the genius Marilyn vos Savant.<br /><br />Suppose we adapt the problem: there are 1,000 doors instead of 3, Monty Hall randomly places a prize behind one of the doors, you select one, and then he opens every other door except one. Suppose you select, say, door 358, and then Monty Hall opens every other door except door 721.&nbsp;<br /><br />Intuitively, it&rsquo;s much, much more probable that door 721 conceals the prize than that the door you happened to select conceals the prize.&nbsp;<br /><br />In fact, the probability is provably 99.9%, and the reason is again because of the likelihoods: if door 721 conceals the prize, there&rsquo;s a 100% likelihood Monty Hall would open every other unselected door except door 721, so <em>P(721 is not opened|door 721 conceals the prize)</em>=100%. But if the door you selected conceals the prize, then there&rsquo;s a 1/999 likelihood that Monty Hall would open every door except 721 &mdash;simply because he had 999 other doors he could not have opened, and so <em>P(721 is not opened|door 358 conceals the prize)</em>=1/999.&nbsp;<br /><br />The point is that this adaptation and others like it show that as we change the number of doors&mdash;and consequently the relevant likelihoods&mdash;we can see how this affects the probability of the relevant hypotheses. And if we decrease the number of doors enough, we will eventually get to just three doors and the 2/3 probability that door B conceals the prize.&nbsp;<br /><br /><br /><em>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; Explanation #3: The Mental Simulations Approach</em><br /><br />A third way to explain the problem is with what I call the <em>mental simulations approach</em> in the <a href="https://www.cambridge.org/core/journals/judgment-and-decision-making/article/likelihood-neglect-bias-and-the-mental-simulations-approach-an-illustration-using-the-old-and-new-monty-hall-problems/B80822FC020A13C4A83E293C6120492E" target="_blank">recent article</a>. This approach involves running a number of mental simulations of the Monty Hall problem and then proportioning the mental simulations by the relevant probabilities.&nbsp;<br /><br />So suppose we first run, say, 30 mental simulations&mdash;that is, we imagine that the Monty Hall problem happens 30 times. And since there is a 1/3 probability that a given door conceals the prize at the beginning, we make it so 1/3 or 10/30 simulations are those where a given door conceals the prize. Then, since the likelihoods mean door C is opened 50% of the time when door A conceals the prize and 100% of the time when door B conceals the prize, we make it so Monty Hall opens door C in 50% and 100% of those simulations respectively. We could depict this as follows:</div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/ms_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div class="paragraph">What these mental simulations show is that 10 out of 15 times, or 2 out of 3 times, when door C is opened, it is because door B conceals the prize, so door B conceals the prize with a probability of 2/3.<br /><br />The experiments further showed that training in the mental simulations approach caused 32% or 16 of 50 participants in one group of the experiment to get the right probabilities for the Monty Hall problem. Of course, this isn&rsquo;t perfect, but it&rsquo;s at least better than the other group where none of the participants got the correct probabilities.<br />&#8203;<br /><br /><em>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; Explanation #4: Computer Simulations</em><br /><br />Another explanation of the problem is actually to just simulate the problem many times using a computer. I provide some code and instructions for how you can do this yourself in Appendix C of the paper <a href="https://static.cambridge.org/content/id/urn%3Acambridge.org%3Aid%3Aarticle%3AS1930297524000081/resource/name/S1930297524000081sup001.zip" target="_blank">here</a>. If you run the simulations, you can see that door B will conceal the prize 2 out of 3 times in the original Monty Hall problem and 10 out of 11 times in the new Monty Hall problem. Apparently, it was only after seeing computer simulations like these that the famous mathematician Paul Erd&#337;s finally accepted the 2/3 solution to the problem and accepted that his initial 1/2 intuition was incorrect.&nbsp;<br /><br /></div>]]></content:encoded></item><item><title><![CDATA[How can we measure the accuracy of judgments and determine which ones to trust?]]></title><link><![CDATA[https://www.johnwilcox.org/johns-blog/how-can-we-measure-the-accuracy-of-judgments-and-determine-which-ones-to-trust]]></link><comments><![CDATA[https://www.johnwilcox.org/johns-blog/how-can-we-measure-the-accuracy-of-judgments-and-determine-which-ones-to-trust#comments]]></comments><pubDate>Sat, 03 Aug 2024 03:05:21 GMT</pubDate><category><![CDATA[Calibrationism]]></category><category><![CDATA[Judgment accuracy]]></category><category><![CDATA[Probability and statistics]]></category><guid isPermaLink="false">https://www.johnwilcox.org/johns-blog/how-can-we-measure-the-accuracy-of-judgments-and-determine-which-ones-to-trust</guid><description><![CDATA[(Forthcoming in&nbsp;Psychology Today)&#8203;  &#8203;THE TL;DR KEY POINTS  I recently published an argument for calibrationism, the idea that judgments about the world are trustworthy only if&mdash;among other things&mdash;there&rsquo;s evidence that they are produced in ways that are &ldquo;well calibrated&rdquo;A set of judgments are well calibrated just in case they assign, say, 80% probabilities to things which are true 80% of the time, 90% probabilities to things which are true 90% of the  [...] ]]></description><content:encoded><![CDATA[<div class="paragraph"><span style="color:rgb(123, 123, 123)">(Forthcoming in&nbsp;</span><em style="color:rgb(123, 123, 123)">Psychology Today</em><span style="color:rgb(123, 123, 123)">)<br />&#8203;</span><br /></div>  <h2 class="wsite-content-title">&#8203;THE TL;DR KEY POINTS</h2>  <div class="paragraph"><ul><li><em><strong>I recently published an argument for <a href="https://link.springer.com/article/10.1007/s13164-024-00724-1" target="_blank">calibrationism</a>, the idea that judgments about the world are trustworthy only if&mdash;among other things&mdash;there&rsquo;s evidence that they are produced in ways that are &ldquo;well calibrated&rdquo;</strong></em></li><li><em><strong>A set of judgments are well calibrated just in case they assign, say, 80% probabilities to things which are true 80% of the time, 90% probabilities to things which are true 90% of the time and so on</strong></em></li><li><em><strong>In this post, I provide a tool for measuring calibration, and I share some ideas about what to do and what not to do when measuring calibration, using my own experience as an example&nbsp;</strong></em></li></ul></div>  <div><div style="height: 20px; overflow: hidden; width: 100%;"></div> <hr class="styled-hr" style="width:100%;"></hr> <div style="height: 20px; overflow: hidden; width: 100%;"></div></div>  <div class="paragraph" style="text-align:left;"><strong><font color="#2a2a2a"><font size="5">PRELIMINARY CLARIFICATION: WHAT IS CALIBRATION AND CALIBRATIONISM?</font></font></strong></div>  <div class="paragraph">&#8203;<br />&#8203;As I mentioned <a href="https://www.johnwilcox.org/johns-blog/when-should-we-trust-judgments-from-ourselves-or-others-the-calibrationist-answer" target="_blank">elsewhere</a>, I recently published <a href="https://link.springer.com/article/10.1007/s13164-024-00724-1" target="_blank">a paper</a> arguing for calibrationism, the idea that judgments of probability are trustworthy only if there&rsquo;s evidence they are produced in ways that are calibrated&mdash;that is, only if there is evidence that the things one assigns probabilities of, say, 90% to happen approximately 90% of the time.<br /><br />Below is an example of such evidence; it is a graph which depicts the calibration of a forecaster from the Good Judgment project&mdash;user 3559:</div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/picture1-graph_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div class="paragraph">&#8203;The graph shows how often the things they assign probabilities to turn out to be true. For example, the top right dot represents all the unique events which they assigned a probability of around 97.5% to before they did or didn&rsquo;t occur: that Mozambique would experience an onset of insurgency between October 2013 and March 2014, that France would deliver a Mistral-class ship to a particular country before January 1st, 2015 and so on for 17 other events. Now, out of all of these 19 events which they assigned a probability of about 97%, it turns out that about 95% of those events occurred. Likewise, if you look at all the events this person assigned a probability of approximately 0%, it turns out that about 0% of those events occurred.&nbsp;<br /><br />However, not all people are like this, below is a particular individual, user 4566, who assigned probabilities of around 97% to things which were true merely 21% of time, such as Chad experiencing insurgency by March 2014 and so on.</div>  <div>  <!--BLOG_SUMMARY_END--></div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/poor-calibration_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div class="paragraph">This graph shows that this person is poorly calibrated.&nbsp;<br /><br />Studies have shown people can be more or less calibrated for a range of <a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">domains</a>, including medical diagnoses and prognoses, general knowledge about past or current affairs, geopolitical topic&mdash;virtually anything.<br /><br />Calibrationism then says that judgments of probability are trustworthy only if we have evidence that they are produced by ways of thinking that are well calibrated&mdash;that is, by ways of thinking that look more like the first forecaster&rsquo;s calibration graph than like the second forecaster&rsquo;s calibration graph.&nbsp;<br /><br />A number of studies suggest humans can be miscalibrated to some extent. For example, <a href="https://press.princeton.edu/books/hardcover/9780691178288/expert-political-judgment" target="_blank">Philip Tetlock</a> found that some particular experts assigned 0% probabilities to events that actually did happen 19% of the time. <a href="https://psycnet.apa.org/record/2011-15298-010" target="_blank">Another study</a> found that some students assigned probabilities of about 95% to things that we true only 73% to 87% of the time (depending on which university these particular students were from). A variety of other studies provide evidence of miscalibration in other contexts, as discussed in my book <em><a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">Human Judgment</a>.<br />&#8203;</em></div>  <div class="paragraph"><strong><font size="5" color="#2a2a2a">HOW TO MEASURE ACCURACY AND CALIBRATION</font></strong></div>  <div class="paragraph"><br />&#8203;&#8203;So, the evidence suggests not only that miscalibration is fairly common; what&rsquo;s more, various studies also suggest that we can be unaware of just how miscalibrated or inaccurate we are.<br /><br />Because of this, it is worth gathering evidence about how calibrated the judgments are that we rely on, and calibrationism says this is necessary before we can fully trust them.<br /><br />Gathering this is often entirely possible, though, and the purpose of this post is to share some thoughts about how this is so.<br /><br />For a start, we can plug some values into a spreadsheet. Below is an adaptation from a spreadsheet which I&rsquo;ve used to track my calibration for almost a year now and which can be downloaded <a href="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/accuracy_tracker_-_online.xlsx" target="_blank">here</a>:</div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/picture5_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div class="paragraph">&#8203;It might seem odd to do this, but I don&rsquo;t think it&rsquo;s irrational. I find myself often making probability judgments about the world, and I found it both fun and insightful to take a moment to make this tool to systematically track the accuracy of the judgments which I make regardless.&nbsp; (Plus, as a cognitive scientist who specializes in judgment accuracy, it makes sense to do something like this since studying this kind of thing is literally my job!)<br /><br />In the spreadsheet above, I have anonymized the content so it does not concern other people, but the topics include a broad variety of things: examples include past or current medical diagnoses and prognoses, interpersonal relationships with friends or others, career developments, events in international affairs and many other things. In that sense, it&rsquo;s a truly &ldquo;domain general&rdquo; set of judgments, to use a term from my book on <em><a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">Human Judgment</a></em>. Some aspects of the spreadsheet are self-explanatory (e.g. the &ldquo;Topic&rdquo;, &ldquo;Description&rdquo; and &ldquo;Rationale&rdquo; columns) while others are explained elsewhere (e.g. &ldquo;Brier&rdquo; and &ldquo;Resolution&rdquo; scores, both of which are discussed in the second chapter of my book <a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">here</a>).<br /><br />What matters, though, is that the spreadsheet automatically produces a table and graph to depict the calibration of a set of judgments as follows (note errors like &ldquo;#DIV/0&rdquo; show up in the categories where I haven&rsquo;t made any judgments*):</div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/picture6_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/picture7_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div class="paragraph">Like the other calibration graphs, the <em>x</em>-axis shows the probabilities I assign to things and the <em>y</em>-axis shows the frequency with which those things turn out to be true. So far, the judgments are pretty well calibrated for some categories: if you take the things I have been around 97.5% confident in, for example, about 92% of those things are true (although I will say more on this shortly), while the things I have assigned 70% probabilities to are true about 71% of the time.&nbsp; Other categories may appear less well calibrated: for example, out of the things I assign 80% to, all of those things are true, thus possibly indicating a degree of underconfidence.&nbsp;</div>  <div class="paragraph" style="text-align:left;"><br /><strong><font size="5">THINGS TO PAY ATTENTION TO WHEN MEASURING CALIBRATION<br />&#8203;</font></strong><br /></div>  <div class="paragraph">However, the statistics for some of the categories may be misleading, and so there are a few things to pay attention to when measuring calibration&mdash;there are several desiderata or ideal features, we might say.<br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; &nbsp; &nbsp;&nbsp;</em><em>Desideratum #1: Multiplicity</em><br /><br />One of which is what I call <em>multiplicity</em>&mdash;which concerns the number of judgments we have. When measuring calibration, we often need multiple judgments that can be assessed for accuracy.&nbsp;<br /><br />For instance, even a person who is perfectly calibrated will assign probabilities of 90% to things which are false 10% of the time, and if we only have one of their 90%-judgments for an outcome which happens to be false, then they might look terribly miscalibrated when they are actually not.&nbsp;<br /><br />For example, in the spreadsheet above, I assigned a 95% probability to moving into a particular apartment ahead of time; then the day to move in came, I was on my way to get the keys, and then the apartment-owner said an issue came up and I couldn&rsquo;t move in. That was the one thing I assigned a high probability to in that category which turned out false, and if we focus on just that instance which (I would think) is objectively improbable, I might look much more miscalibrated than I actually am.<br /><br />More generally, the so-called &ldquo;law of large numbers&rdquo; implies that if, say, the things we are 90% confident in would be true 90% of the time, then our actual judgments will likely converge to this proportion&mdash;but only with a sufficiently large number of judgments. It&rsquo;s like how the probability of a coin landing heads is 50%, but the actual numbers reflect this only with multiple coin flips: three coin flips might land heads 100% of the time, or 66% of the time, but many more flips will eventually reveal the true probability of 50%.&nbsp;<br /><br />Consequently, the number of judgments in a given category is depicted in the calibration graph. There are 13 judgments in the 97.5% category and only 2 in the 60% category, meaning that the observed calibration is much more likely to reflect the true calibration for the former category than for the latter.<br /><br /><em style="color:rgb(123, 123, 123)">&nbsp; &nbsp; &nbsp; &nbsp; &nbsp;</em><em style="color:rgb(123, 123, 123)">&nbsp;</em><em>Desideratum #2: Independence</em><br /><br />However, observed calibration is more likely to reflect one&rsquo;s real underlying accuracy when another desideratum is met: namely, that the things one makes judgments about are probabilistically independent of each other. Suppose, for example, that the probability the Republican candidate will win the next US presidential election is 70%. Then, if they have a dataset containing only 100 of the same judgments that are each 70% confident in his success, then either all of the predictions are true or none of them are, since they all concern the exact same thing.<br /><br />More generally, then, a set of judgments could look unduly inaccurate merely because the judgments are dependent on each other, either because they are about the same topic (e.g. repeating the same forecast in a forecasting tournament), or because they are about closely related topics which affect each other (e.g. the number of COVID deaths and the number of COVID infections in a country).<br /><br />One way to guarantee the right kind of independence, then, is to make sure that the judgments each concern an independent topic: you might have one judgment about the US presidential election, another about the culture of Mexico, another about a question of geography or whatever else one might care about.<br /><br /><em>&nbsp; &nbsp; &nbsp; &nbsp; &nbsp; Desideratum #3: Resolvability</em><br /><br />But beyond that, however, we would ideally assess accuracy using questions that are "resolvable", meaning we can know with virtual certainty what the correct answer is. Sometimes we can do this, like when predicting a future outcome which will eventually be knowable or when using things like DNA evidence to figure out what happened in the past.<br /><br />But sometimes the questions we are concerned with are not always resolvable. For example, one study assessed the accuracy of judgments from experts vs. non-experts by comparing the two groups&rsquo; judgments to estimates from other experts.&nbsp; But if the &ldquo;other experts&rdquo; have inaccurate judgments, as experts sometimes do, then this would give a skewed and possibly biased picture of judgment accuracy.<br /><br />In any case, the upshot is that it&rsquo;s very feasible to measure the calibration of multiple, independent and resolvable judgments using the free spreadsheet tool above. This could potentially help us determine both the trustworthiness of those judgments and<span style="color:rgb(123, 123, 123)">&mdash;by extrapolation--</span>those in other contexts where the true outcomes might not be knowable.<br /><br /><br /><br /><em>*Footnote:&nbsp;<span style="color:rgb(123, 123, 123)">In my calibration graph above, I've "flipped" probabilities like 10% in false propositions to become 90% in true propositions so I have a larger sample of judgments for the categories above 50%.</span><br /></em><br /></div>]]></content:encoded></item><item><title><![CDATA[When should we trust the judgments from ourselves or others? The calibrationist answer]]></title><link><![CDATA[https://www.johnwilcox.org/johns-blog/when-should-we-trust-judgments-from-ourselves-or-others-the-calibrationist-answer]]></link><comments><![CDATA[https://www.johnwilcox.org/johns-blog/when-should-we-trust-judgments-from-ourselves-or-others-the-calibrationist-answer#comments]]></comments><pubDate>Sat, 03 Aug 2024 02:40:51 GMT</pubDate><category><![CDATA[Calibrationism]]></category><category><![CDATA[Probability and statistics]]></category><category><![CDATA[Rationality]]></category><guid isPermaLink="false">https://www.johnwilcox.org/johns-blog/when-should-we-trust-judgments-from-ourselves-or-others-the-calibrationist-answer</guid><description><![CDATA[(Forthcoming in&nbsp;Psychology Today)&#8203;  &#8203;THE TL;DR KEY POINTS  We all make judgments of probability and use these to inform our decision-makingBut it is not obvious which judgments to trust, and bad outcomes occur if we get it wrong&mdash;such as fatal misdiagnoses or false death sentence convictionsHow do we determine which judgments to trust&mdash;both from ourselves or others?I recently argued for inclusive calibrationism, which&nbsp;gives a two-part answer to this questionThe fi [...] ]]></description><content:encoded><![CDATA[<div class="paragraph"><span style="color:rgb(123, 123, 123)">(Forthcoming in&nbsp;</span><em style="color:rgb(123, 123, 123)">Psychology Today</em><span style="color:rgb(123, 123, 123)">)<br />&#8203;</span><br /></div>  <h2 class="wsite-content-title">&#8203;THE TL;DR KEY POINTS</h2>  <div class="paragraph"><ul><li><strong><em>We all make judgments of probability and use these to inform our decision-making</em></strong></li><li><strong><em>But it is not obvious which judgments to trust, and bad outcomes occur if we get it wrong&mdash;such as fatal misdiagnoses or false death sentence convictions</em></strong></li><li><strong><em>How do we determine which judgments to trust&mdash;both from ourselves or others?</em></strong></li><li><strong><em>I recently argued for <a href="https://osf.io/preprints/psyarxiv/vm34x" target="_blank">inclusive calibrationism</a>, which&nbsp;gives a two-part answer to this question</em></strong></li><li><strong><em>The first part says judgments of probability are trustworthy only if there&rsquo;s evidence they are produced in ways that are "well calibrated"&mdash;that is, only if there is evidence that the things one assigns probabilities of, say, 90% to happen approximately 90% of the time</em></strong></li><li><strong><em>The second part says that judgments of probability are trustworthy only if they are also inclusive of all the relevant evidence</em></strong></li><li><strong><em>This blogpost then shares some ideas for implementing calibrationism, such as measuring calibration and creating evidence checklists to figure out how inclusive a judgment is</em></strong></li></ul></div>  <div><div style="height: 20px; overflow: hidden; width: 100%;"></div> <hr class="styled-hr" style="width:100%;"></hr> <div style="height: 20px; overflow: hidden; width: 100%;"></div></div>  <div class="paragraph"><span style="color:rgb(123, 123, 123)">&#8203;</span><strong style="color:rgb(123, 123, 123)"><font color="#2a2a2a" size="5">THE IMPORTANCE OF TRUSTWORTHY JUDGMENTS</font></strong><br /><br />We all make judgments of probability and depend on them for our decision-making.<br /><br />However, it is not always obvious which judgments to trust, especially since a range of studies suggest these judgments can sometimes be more inaccurate than we might hope or expect. For example, <a href="https://www.pnas.org/doi/abs/10.1073/pnas.1306417111" target="_blank">scholars</a> have argued at least 4% of death sentence convictions in the US are false convictions, that <a href="https://jamanetwork.com/journals/jama/article-abstract/1845204" target="_blank">tens</a> or even <a href="https://europepmc.org/article/NBK/nbk588118" target="_blank">hundreds of thousands</a> of Americans die of misdiagnoses each year and that sometimes <a href="https://press.princeton.edu/books/hardcover/9780691178288/expert-political-judgment" target="_blank">experts can be 100% sure of predictions</a> which turn out to be false 19% of the time. So we want trustworthy judgments, or else bad outcomes can occur.<br /><br />How do, then, can we determine which judgments to trust&mdash;either from ourselves or others? In a paper recently published <a href="https://link.springer.com/article/10.1007/s13164-024-00724-1" target="_blank">here</a> and freely available <a href="https://osf.io/preprints/psyarxiv/vm34x" target="_blank">here</a>, I argue for an answer called &ldquo;inclusive calibrationism&rdquo;&mdash;or just &ldquo;calibrationism&rdquo; for short. Calibrationism says trustworthiness requires two ingredients&mdash;calibration and inclusivity.&nbsp;<br />&#8203;</div>  <div>  <!--BLOG_SUMMARY_END--></div>  <div class="paragraph" style="text-align:left;"><strong style="color:rgb(123, 123, 123)"><font color="#2a2a2a" size="5">THE FIRST INGREDIENT OF TRUSTWORTHINESS: CALIBRATION</font></strong><br /><strong><span><span style="color:rgb(0, 0, 0)"><font size="4">&#8203;</font></span></span></strong></div>  <div class="paragraph">&#8203;&#8203;&#8203;The calibrationist part of &ldquo;inclusive calibrationism&rdquo; says that judgments of probability are trustworthy only if they are produced by ways of thinking that have evidence of their calibration. Here, &ldquo;calibration&rdquo; is a technical term that refers to whether the probabilities we assign to things correspond to the frequency with which those things are true. Let us consider this with an example.&nbsp;<br /><br />Below is a graph which depicts the calibration of a forecaster from the Good Judgment project&mdash;user 3559:</div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/picture1-graph_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div class="paragraph">&#8203;The graph shows how often the things they assign probabilities to turn out to be true. For example, the top right dot represents all the unique events which they assigned a probability of around 97.5% to before they did or didn&rsquo;t occur: that Mozambique would experience an onset of insurgency between October 2013 and March 2014, that France would deliver a Mistral-class ship to a particular country before January 1st, 2015 and so on for 17 other events. Now, out of all of these 19 events which they assigned a probability of about 97%, it turns out that about 95% of those events occurred. Likewise, if you look at all the events this person assigned a probability of approximately 0%, it turns out that about 0% of those events occurred.&nbsp;<br /><br />In this case, this person has a relatively good track record of calibration because they think in ways which mean that the probabilities that they assign to things correspond (roughly) to the frequency with which those things are true.&nbsp;<br /><br />And this is the case even when those things concern &ldquo;unique&rdquo; events which we might have thought we couldn&rsquo;t assign probabilities too. After all, a particular insurgency either would or would not occur, it&rsquo;s a unique event, and so it&rsquo;s not obvious we can assign it a numerically precise number, like 67%, that&mdash;as it turns out&mdash;reflects the objective odds of its occurrence. But it turns out that we can&mdash;at least if we think in the right ways.<br /><br />Yet not all of us are so well calibrated. Below is a particular individual, user 4566, who assigned probabilities of around 97% to things which were true merely 21% of time, such as Chad experiencing insurgency by March 2014 and so on.</div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/picture2_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div class="paragraph">&#8203;Studies have shown people can be more or less calibrated for a range of domains, including <a href="https://psycnet.apa.org/record/2003-02858-039" target="_blank">medical diagnoses</a> and prognoses, general knowledge about <a href="https://psycnet.apa.org/record/2011-15298-010" target="_blank">past or current affairs</a>, <a href="https://press.princeton.edu/books/hardcover/9780691178288/expert-political-judgment" target="_blank">geopolitical topics</a>&mdash;virtually anything.<br /><br />Calibrationism then says that judgments of probability are trustworthy only if we have evidence that they are produced by ways of thinking that are well calibrated&mdash;that is, by ways of thinking that look more like the first forecaster&rsquo;s calibration graph than like the second forecaster&rsquo;s calibration graph.&nbsp;<br /><br />Of course, for many kinds of judgments, we lack strong evidence of calibration, and we then have little grounds for unquestioning trust in those judgements of probability. This is true in some parts of medicine or law, for instance, where we have some evidence of inaccurate judgments which can lead to misdiagnoses or false convictions at least some of the time, even if we have perfectly good judgments the rest of the time.&nbsp;<br /><br />But in other contexts, we simply lack evidence either way: our judgments could be as calibrated as the first forecaster or as miscalibrated as the second, and we have no good reasons to tell firmly either way.&nbsp;<br /><br />The calibrationist implication is that in some domains, then, we need to measure and possibly improve our calibration before we can fully trust the judgments of probability which we have in that domain.&nbsp;<br /><br />The good news, however, is that this is possible, as I discuss in my book <a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">elsewhere</a>. For example, some evidence (e.g.&nbsp;<a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3779404" target="_blank">here</a> and <a href="https://psycnet.apa.org/record/2015-00693-001" target="_blank">here</a>) suggests we can improve our calibration by drawing on information about statistics or the frequency with which things have happened in the past. For instance, we can better predict a recession in the future if we look at the proportion of the time that recessions have occurred in similar situations in the past. We can similarly use statistics like these to determine the probabilities of election outcomes, wars, a patient having a disease and even more mundane outcomes like whether someone has a crush on you.<br /><br />The other good news is that some people are well calibrated, meaning that they can provide us with trustworthy judgments about the world. For example, take user 5265 from the Good Judgment&rsquo;s forecasting tournament. In the first year of the tournament, their judgments were well calibrated, as the below graph depicts:</div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/picture3_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div class="paragraph">&#8203;If we were in the second year of the Good Judgment tournament, and we were about to ask this person another series of questions, we could then have similarly inferred the calibration and trustworthiness of their judgments about the future. And in fact, that is exactly what we see when we look at their track record of calibration for the second year of the tournament below:</div>  <div><div class="wsite-image wsite-image-border-none " style="padding-top:10px;padding-bottom:10px;margin-left:0;margin-right:0;text-align:center"> <a> <img src="https://www.johnwilcox.org/uploads/6/3/9/4/63943309/picture4_orig.png" alt="Picture" style="width:auto;max-width:100%" /> </a> <div style="display:block;font-size:90%"></div> </div></div>  <div class="paragraph">&#8203;More generally, <a href="https://link.springer.com/chapter/10.1007/978-3-031-30085-1_6" target="_blank">the evidence demonstrates</a> track records of accuracy are the biggest indicator of someone&rsquo;s accuracy in other contexts&mdash;and better indicators than education level, age, experience or anything else that has been scientifically tested.&nbsp;<br /><br />So track records of calibration are one important ingredient of trustworthiness, but it&rsquo;s not the only one.&nbsp;<br />&#8203;</div>  <div class="paragraph" style="text-align:left;"><strong><font color="#2a2a2a" size="5">THE SECOND INGREDIENT OF TRUSTWORTHINESS: INCLUSIVITY</font></strong><br /></div>  <div class="paragraph"><br />&#8203;Another important ingredient is inclusivity: that is, the extent to which we consider all the evidence which is relevant.&nbsp; After all, calibration isn&rsquo;t everything we care about, since someone&nbsp; could be perfectly well calibrated merely by assigning 50% probabilities to a series of &ldquo;yes/no&rdquo; questions. Additionally, <a href="https://www.cambridge.org/core/journals/judgment-and-decision-making/article/role-of-actively-openminded-thinking-in-information-acquisition-accuracy-and-calibration/1D78BE16863F3F6B2D1C8B5307C9C3B3" target="_blank">some evidence</a> suggests people form more accurate judgments when they include more evidence than others&mdash;and it&rsquo;s obvious how this can improve accuracy when, for example, including DNA evidence which vindicates otherwise convicted defendants in law.<br /><br />What we also care about, then, is whether judgments of probability are informative in the sense that they tell us whether something is true or not in a particular case. This in turn is largely a matter of including evidence. For example, one could be well calibrated by assigning 50% probabilities for the &ldquo;yes/no&rdquo; questions, but they would likely be omitting relevant evidence and not saying anything particularly informative.<br /><br />Calibrationism then says that in order for judgments to be trustworthy, they must also include all the evidence which we regard as relevant.<br />&#8203;</div>  <div class="paragraph"><font color="#2a2a2a" size="5"><strong>GETTING PRACTICAL: <br />&#8203;HOW TO IMPLEMENT CALIBRATIONISM</strong></font></div>  <div class="paragraph"><br />&#8203;So that&rsquo;s calibrationism in a rough nutshell: our judgments are trustworthy to the extent we have evidence that A) they are produced in ways that are well calibrated and B) they are inclusive of all the relevant evidence. What, then, are the implications of this? Four come to mind:<br />&#8203;<ol><li>We should measure calibration to determine which judgments are trustworthy</li><li>We should trust individuals who have evidence of their calibration&nbsp;</li><li>But only if they are inclusive of all the evidence</li><li>And where necessary, we should aim to improve our calibration</li></ol><br />Practically, I provide more ideas about how we can do these things elsewhere. For example, one can measure calibration by plugging some judgments into a spreadsheet template, as I discuss <a href="https://www.johnwilcox.org/johns-blog/how-can-we-measure-the-accuracy-of-judgments-and-determine-which-ones-to-trust" target="_blank">here</a>. Calibration can also be improved with recommendations which I discuss <a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">here</a>. Lastly, if we want to assess the inclusivity and trustworthiness of someone&rsquo;s thinking, we can list the evidence which we think is relevant, ask them questions about their reactions to each item, and if their responses seem to reflect calibrated engagement with all of the evidence, then we can trust their judgments.<br /><br />Put simply, to determine which judgments to trust, we might want to see more calibration graphs and evidence checklists to determine calibration and inclusivity&mdash;at least when the stakes are high. This might help make a world with fewer fatal misdiagnoses, false criminal convictions and other expressions of inaccuracy that compromise the functioning and well-being of our societies.</div>]]></content:encoded></item><item><title><![CDATA[New book, "Human Judgment", is now published]]></title><link><![CDATA[https://www.johnwilcox.org/johns-blog/new-book-human-judgment-is-now-published]]></link><comments><![CDATA[https://www.johnwilcox.org/johns-blog/new-book-human-judgment-is-now-published#comments]]></comments><pubDate>Mon, 02 Jan 2023 11:22:26 GMT</pubDate><category><![CDATA[Heuristics and biases]]></category><category><![CDATA[Probability and statistics]]></category><guid isPermaLink="false">https://www.johnwilcox.org/johns-blog/new-book-human-judgment-is-now-published</guid><description><![CDATA[&#8203;THE TL;DR KEY POINTS  We all make judgments, and our important life decisions depend&nbsp;on them--but how good are those judgments?My new book "Human Judgment: How Accurate is It, and How Can it Get Better?" investigates this topic (it's now available to purchase here)It has two somewhat newsworthy items: one bad, the other goodThe bad news is that&nbsp;science suggests humans are often much more inaccurate than we might hope or expect--for example,&nbsp;thousands die each year because o [...] ]]></description><content:encoded><![CDATA[<h2 class="wsite-content-title">&#8203;THE TL;DR KEY POINTS</h2>  <div class="paragraph"><ul><li><em><strong>We all make judgments, and our important life decisions depend&nbsp;on them</strong></em><span style="color:rgb(123, 123, 123)">--<strong><em>but how good are those judgments?</em></strong></span></li><li><strong><em>My new book "Human Judgment: How Accurate is It, and How Can it Get Better?" investigates this topic (it's now available to purchase <a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">here</a>)</em></strong></li><li><strong><em>It has two somewhat newsworthy items: one bad, the other good</em></strong></li><li><strong><em>The bad news is that&nbsp;science suggests humans are often much more inaccurate than we might hope or expect</em></strong><span style="color:rgb(123, 123, 123)">--</span><strong><em>for example,&nbsp;thousands die each year because of&nbsp;misdiagnoses and&nbsp;false death sentence convictions</em></strong></li><li><strong><em>The good news is that&nbsp;science suggests a number&nbsp;of concrete ways to measure and improve the accuracy of our judgments</em></strong></li><li><strong><em>The book then outlines and summarizes those recommendations</em></strong></li></ul></div>  <div><div style="height: 20px; overflow: hidden; width: 100%;"></div> <hr class="styled-hr" style="width:100%;"></hr> <div style="height: 20px; overflow: hidden; width: 100%;"></div></div>  <div class="paragraph">We all make countless judgments, and our important life decisions depend on them.&nbsp;&nbsp;<br /><br />My new book, &ldquo;Human Judgment&rdquo;, investigates these judgments, and it is now available to purchase online <a href="https://link.springer.com/book/10.1007/978-3-031-19205-0" target="_blank">here</a>.<br /><br />The book concerns two topics to do with human judgment, as implied by the subtitle: How accurate is it, and how can it get better?<br /><br />It has two somewhat newsworthy items, one bad and the other good.<br /><br />The bad news is that the science suggests that human judgment is often much more inaccurate than we might hope or expect. For example, some researchers estimated as many as 40,000 to 80,000 US citizens will die because of preventable misdiagnoses&mdash;and that&rsquo;s <em>each year</em>. If they are right, that&rsquo;s a yearly death toll at least 13 times higher than the September 11th terrorist attacks. Unfortunately, medicine is not unique too: judgmental inaccuracy can afflict a number of other areas in society as well. As another example, some researchers estimate at least 4.1% of death sentence convictions in the US are actually false convictions; this implies that some people are trialed, convicted and executed for horrific crimes that they never actually committed. So that is a few of numerous studies painting a less than ideal picture of human judgment: we make inaccurate judgments about medical diagnoses, about criminal convictions and about a number of other areas.</div>  <div>  <!--BLOG_SUMMARY_END--></div>  <div class="paragraph"><br /><span style="color:rgb(123, 123, 123)">Of course, we are not always so inaccurate. We often judge with perfect accuracy where we live or whether humans need oxygen to survive, to take a couple of many mundane examples. To that extent, the book espouses a&nbsp;</span><em style="color:rgb(123, 123, 123)">context dependent model of human accuracy</em><span style="color:rgb(123, 123, 123)">: how accurate we are simply depends on the context, thus prohibiting unqualified generalizations.</span><br /><br /><span style="color:rgb(123, 123, 123)">Regardless, we are substantially inaccurate in an important remainder of other contexts, and then there are many more contexts where we simply do not know how accurate humans are.</span><br /><br /><span style="color:rgb(123, 123, 123)">The good news, however, is that science also suggests some ways to measure and improve accuracy. The hope, then, is that the book can motivate society to capitalize on those ways where possible, and it gives some recommendations to that extent.</span><br /><br /><span style="color:rgb(123, 123, 123)">Most of the recommendations are informed by pioneering research funded by the US intelligence community. US intelligence has funded some of the most cutting-edge research in improving the accuracy of human judgment. The book then describes a lot of that research. However, often the research concerns specifically&nbsp;</span><em style="color:rgb(123, 123, 123)">geopolitical topics</em><span style="color:rgb(123, 123, 123)">--</span><span style="color:rgb(123, 123, 123)">topics like the outcomes of wars, elections and the like. But that said, the research also supports a set of generalizable recommendations for improving human judgment&mdash;or so I argue in the book.</span><br /><br /><span style="color:rgb(123, 123, 123)">The ultimate long-term aim of the book is that society will use these recommendations to improve our judgments, our decision-making and ultimately our lives&mdash;thereby reducing misdiagnoses, false convictions and many other expressions of judgmental inaccuracy that severely compromise human well-being. Fingers crossed it eventually achieves its aims!</span></div>  <div> 	<form enctype="multipart/form-data" action="//www.weebly.com/weebly/apps/formSubmit.php" method="POST" id="form-564518087779446225"> 		<div id="564518087779446225-form-parent" class="wsite-form-container" 				 style="margin-top:10px;"> 			<ul class="formlist" id="564518087779446225-form-list"> 				<h2 class="wsite-content-title">Critical feedback survey</h2>  <label class="wsite-form-label wsite-form-fields-required-label"><span class="form-required">*</span> Indicates required field</label><div><div class="wsite-form-field" style="margin:5px 0px 5px 0px;"> 				<label class="wsite-form-label" for="input-666660850360175062">What is your feedback (positive or negative) about the book "Human judgment"? All of your opinions are valued! <span class="form-required">*</span></label> 				<div class="wsite-form-input-container"> 					<textarea aria-required="true" id="input-666660850360175062" class="wsite-form-input wsite-input wsite-input-width-500px" name="_u666660850360175062" style="height: 200px"></textarea> 				</div> 				<div id="instructions-666660850360175062" class="wsite-form-instructions" style="display:none;"></div> 			</div></div> 			</ul> 			 		</div> 		<div style="display:none; visibility:hidden;"> 			<input type="hidden" name="weebly_subject" /> 		</div> 		<div style="text-align:left; margin-top:10px; margin-bottom:10px;"> 			<input type="hidden" name="form_version" value="2" /> 			<input type="hidden" name="weebly_approved" id="weebly-approved" value="approved" /> 			<input type="hidden" name="ucfid" value="564518087779446225" /> 			<input type="hidden" name="recaptcha_token"/> 			<input type="submit" role="button" aria-label="Submit feedback" value="Submit feedback" style="position:absolute;top:0;left:-9999px;width:1px;height:1px" /> 			<a class="wsite-button"> 				<span class="wsite-button-inner">Submit feedback</span> 			</a> 		</div> 	</form> 	<div id="g-recaptcha-564518087779446225" class="recaptcha" data-size="invisible" data-recaptcha="0" data-sitekey="6Ldf5h8UAAAAAJFJhN6x2OfZqBvANPQcnPa8eb1C"></div>    </div>]]></content:encoded></item></channel></rss>