<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Statistical Modeling, Causal Inference, and Social Science</title>
	<atom:link href="https://statmodeling.stat.columbia.edu/feed/" rel="self" type="application/rss+xml" />
	<link>https://statmodeling.stat.columbia.edu</link>
	<description></description>
	<lastBuildDate>Sun, 26 Jan 2025 21:29:00 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.4.3</generator>
	<item>
		<title>Reading the referee reports of that retracted paper by the science reformers:  A peek behind the curtain</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/26/reading-the-referee-reports-of-that-retracted-paper-by-the-science-reformers-a-peek-behind-the-curtain/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/26/reading-the-referee-reports-of-that-retracted-paper-by-the-science-reformers-a-peek-behind-the-curtain/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Sun, 26 Jan 2025 14:33:40 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Sociology]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51109</guid>

					<description><![CDATA[One interesting sidelight of the story of a much criticized and finally retracted article on replication in psychology is that we get to read some of the referee reports. These aren&#8217;t the reports on the original article&#8212;I guess those reviews &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/26/reading-the-referee-reports-of-that-retracted-paper-by-the-science-reformers-a-peek-behind-the-curtain/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>One interesting sidelight of the story of a <a href="https://statmodeling.stat.columbia.edu/2024/03/27/the-feel-good-open-science-story-versus-the-preregistration-who-do-you-think-wins/">much criticized</a> and <a href="https://statmodeling.stat.columbia.edu/2024/09/24/well-today-we-find-our-heroes-flying-along-smoothly/">finally retracted</a> article on replication in psychology is that we get to read some of the referee reports.</p>
<p>These aren&#8217;t the reports on the original article&#8212;I guess those reviews were positive, otherwise it wouldn&#8217;t have been published in the first place&#8212;but on a proposed replacement paper; Berna Devezer <a href="https://statmodeling.stat.columbia.edu/2024/09/24/well-today-we-find-our-heroes-flying-along-smoothly/#comment-2380149">explains the background</a>.</p>
<p>The reviews are in <a href="https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Decline-Effect-Appeal-Letterfinal.pdf">this document</a>, which came from <a href="https://osf.io/rnvxk">this</a> public repository.</p>
<p>The first two reviews (by Tal Yarkoni and Daniel Lakens) are really long, and they are both very thoughtful, really the equivalent of publishable articles (or blog posts) on their own.  I guess they knew this was a high-stakes case, so they put in extra effort.</p>
<p>The third review was more cursory, and its summary was, &#8220;This is an important and novel set of findings&#8221; and &#8220;The approach is valid, and the quality of data and of its presentation are good.&#8221;  After reading reviews 1 and 2, I don&#8217;t think reviewer 3&#8217;s assessment is correct&#8212;but I&#8217;m not gonna get all upset at reviewer 3.  I&#8217;ve reviewed hundreds, maybe thousands of papers.  I do <a href="https://statmodeling.stat.columbia.edu/2006/10/25/advice_for_refe/">a quick read</a>, and sometimes I miss the point!  That&#8217;s why a journal will obtain three reviews&#8212;different reviews have different styles, and any given reviewer can make a mistake.</p>
<p>Anyway, for those of you who have never been involved in editing a scholarly journal, it might be interesting for you to read these reviews, just to get a peek inside a system that is usually hidden from outsiders.</p>
<p>Some of the reviewers&#8217; comments are just so on the mark:</p>
<blockquote><p>The paper&#8217;s central claim&#8212;i.e., that a high rate of replication failures is not inevitable if one uses optimal practices—is obviously true, and requires no empirical support. The only way it could be false is if there were no reliable effects at all in social science&#8212;which is on its face an absurd proposition. The authors write that they &#8220;report the results of a prospective replication examining whether low replicability and declining effects are inevitable when using current optimal practices.&#8221; But how could low replicability and declining effects possibly be &#8220;inevitable&#8221;, either with or without optimal practices?</p></blockquote>
<p>Good point!</p>
<blockquote><p>A charitable reading of the authors&#8217; central claim might be that what they are really trying to do is quantitatively estimate the impact of using best practices on the replicability of previous findings. That is, the question is not really whether or not replicability is *achievable* (of course it is), but *under what conditions a certain degree of replicability is achievable*, and *how much of an impact certain procedures seem to make*. If the paper were explicitly framed this way, I think it would be posing an important question. A serious effort to quantify the impact of implementing specific rigorous methodologies on replicability would be a valuable service to the field. However, framing the paper this way would also make it clear that, at least in its current presentation, the study has a number of major design limitations that in my view preclude it from providing an informative answer to the question.</p></blockquote>
<p>Bang!  The reviewer is not trying to be mean; he&#8217;s just stating some truths.</p>
<blockquote><p>The generalization target is unclear. It is never made clear in the manuscript what population of effects the authors take their conclusions to apply to. . . . absent the authors&#8217; disclosure of the processes that led them to select these particular effects for inclusion, it is impossible to determine—even in a ballpark sense—what population of studies the present results are meant to apply to.</p></blockquote>
<p>That sort of thing is often important. It&#8217;s the kind of thing that people can be sloppy about in the writing of the paper and that journals will often just let through because nobody cares about going through and getting everything right.</p>
<blockquote><p>
Inadequate description of effects.
</p></blockquote>
<p>That one&#8217;s huge.  Here&#8217;s an example:</p>
<blockquote><p>For example, for the &#8220;Ads&#8221; study, Table 1 describes the central result as &#8220;Watching a short ad within a soap-operate episode increases one&#8217;s likelihood to recommend and promote the company in the ad&#8221;. This is incorrect in at least two ways. First, the authors did not measure likelihood of recommendation or promotion of companies; they measured *self-reported* likelihoods of these behaviors. That there is at best a very weak relationship between these things should be obvious, or else McDonald&#8217;s would experience a significant boost in revenue every time it aired a single ad on TV, which is obviously not the case (indeed, there is a cottage industry within marketing research questioning whether TV ad campaigns have *any* meaningful effect on sales). Second, the current wording implies that the effect applies to companies in general, when actually the authors only asked about McDonald&#8217;s, and used only a single ad comparison (McDonald&#8217;s vs. Prudential). This design does not license any general conclusions about &#8220;the company in the ad&#8221;; it licenses conclusions only about McDonald&#8217;s.</p></blockquote>
<p>This reminds me of <a href="http://stat.columbia.edu/~gelman/research/published/healing3.pdf">our interrogation</a> of a psychology paper that kept describing things inaccurately.  It can be so hard for authors to just say exactly what they did!</p>
<blockquote><p>Of the &#8220;Cookies&#8221; study, the authors write: &#8220;People will be seen as greedier when they take three of the same kind of (free) cookie than when they take three different (free) cookies&#8221;. This is an inaccurate description, as no cookie-takers were actually observed; participants were asked to *imagine* how they would feel if they observed people taking cookies. A more accurate description would be: &#8220;Participants directed to imagine a specific hypothetical norm-violation scenario rate it as greedier to take three of the same kind of (free) cookie than to take three different (free) cookies.&#8221; </p></blockquote>
<p>You might see this as picky, but <a href="https://statmodeling.stat.columbia.edu/2019/10/07/are-statistical-nitpickers-e-g-kaiser-fung-and-me-getting-the-way-of-progress-or-even-serving-the-forces-of-evil/">I don&#8217;t</a>.  We&#8217;re supposed to be doing science here!  You say what you actually did.  Even though, I know that it&#8217;s absolutely standard practice to not do this (notoriously, &#8220;That a person can, by assuming two simple 1-min poses, embody power and instantly become more powerful has real-world, actionable implications,&#8221; describing a study that had no measures of power, let alone &#8220;instantly becoming more powerful&#8221;).</p>
<blockquote><p>At several points in the paper the authors remark the performed studies with ‘high statistical power’ (e.g., abstract). But this is not true. . . . If this confirmatory study happens to be a study where, as a fluke an effect size was observed that was rather extreme (as should happen, by change, in 16 studies), and the true effect was smaller (as it seems to be, in the replication studies), then this was not a sufficiently powered study. Instead, it *seemed to be* a sufficiently powered study, based on the *observed* effect size. But if we take the four replications as an estimate of the true effect size, the study had low power. Of course, all of this requires some speculation, as we never know the true effect size, but the point is, the authors can not argue the studies all had high power, and the absence of a power analysis (let alone a conservative power analysis, such as a safeguard power analysis) should make the authors even more careful about claiming they had ‘high power’. Power is a curve. . . .</p></blockquote>
<p>That&#8217;s a lot of words, but it&#8217;s understandable.  It can take a lot of words to explain what&#8217;s going on to people who have made a mistake.  And it&#8217;s just a referee report; there&#8217;s no reason the reviewer should put in lots of time to write it crisply.  The point is, yeah, the statement in the article was wrong.  And, again, the reviewer isn&#8217;t being mean, he&#8217;s just telling it like it is.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/26/reading-the-referee-reports-of-that-retracted-paper-by-the-science-reformers-a-peek-behind-the-curtain/feed/</wfw:commentRss>
			<slash:comments>5</slash:comments>
		
		
			</item>
		<item>
		<title>Thou Shalt Not Cheat</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/25/thou-shalt-not-cheat/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/25/thou-shalt-not-cheat/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Sat, 25 Jan 2025 14:40:22 +0000</pubDate>
				<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51122</guid>

					<description><![CDATA[Palko points to this post by Remy Levin and writes: JMR, a top 4 marketing journal, has issued a formal “Expression of Concern” for Mazar, Amir, &#038; Ariely (2008), the infamous Ten Commandments study. We discussed some version of that &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/25/thou-shalt-not-cheat/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Palko points to <a href="https://x.com/RemyLevin/status/1838795608968106406">this post</a> by Remy Levin and writes:</p>
<blockquote><p>JMR, a top 4 marketing journal, has issued a formal “Expression of Concern” for Mazar, Amir, &#038; Ariely (2008), the infamous Ten Commandments study.</p></blockquote>
<p><a href="https://statmodeling.stat.columbia.edu/2018/12/13/oh-hate-work-criticized-case-fails-attempted-replications-original-researchers-dont-even-consider-possibility-maybe-original-work-w/">We discussed some version</a> of that experiment a few years ago.  Lots of colorful characters came up. In addition to TV star psychologist Dan Ariely, there was ultra-connected economist Andrei Shleifer, who was credibly accused of something like embezzlement, and of course there was the big irony that there was <a href="https://statmodeling.stat.columbia.edu/2023/06/21/ted-talking-data-fakers-who-write-books-about-lying-and-rule-breaking-whats-up-with-that/">cheating in a study about cheating</a>.</p>
<p>At that point I would&#8217;ve said that the irony meter had gone off the scale, had it not been for the story of Michael Shermer, the professional skeptic who, in apparent sincerity, shared the story of a haunted radio (yes, <a href="https://statmodeling.stat.columbia.edu/2014/12/12/saying-things-place/">really</a>!).</p>
<p>Amazingly&#8212;or not so amazingly&#8212;there&#8217;s a connection between these two denizens of the irony zone, as Shermer <a href="https://statmodeling.stat.columbia.edu/2024/04/07/evilicious-3-face-the-music/">blurbed</a> Ariely&#8217;s recent book.  Ahhhh, celebrities!</p>
<p>And at this point you probably won&#8217;t be surprised to hear that both Ariely and Shermer have been involved with the <a href="https://www.buzzfeednews.com/article/peteraldhous/jeffrey-epstein-john-brockman-edge-foundation">notorious</a> Edge Foundation.  Funky paper <a href="https://statmodeling.stat.columbia.edu/2024/09/15/shreddergate-a-fascinating-investigation-into-possible-dishonesty-in-a-psychology-experiment/">shredders</a>, magic radios, . . . can&#8217;t get much edgier than that.  No <a href="https://statmodeling.stat.columbia.edu/2019/10/02/schoolmarms-and-lightning-bolts-data-faker-meets-edge-foundation-in-an-unintentional-reveal-of-problems-with-the-great-man-model-of-science/">schoolmarms</a> there!</p>
<p>Anyway, the best comment on that Ten Commandments study came from <a href="https://statmodeling.stat.columbia.edu/2018/12/13/oh-hate-work-criticized-case-fails-attempted-replications-original-researchers-dont-even-consider-possibility-maybe-original-work-w/#comment-926306">Jonathan, back in 2018</a>:</p>
<blockquote><p>Where in the 10 Commandments does it say ‘don’t cheat on a matrix test?’ It isn’t murder. It isn’t sex with a married woman (because that’s what it actually says is outlawed – because that confuses parentage and causes violence). It isn’t coveting. It could be lying but was this within the confines of a honor code with meaning, like part of a college commitment, or was it just an agreement to take this test? Does the lying gain someone advantages? The point of the ‘lying’ commandment is bearing false witness, and that would be stretch. Aren’t the 10 Commandments more a lesson that little crap like this doesn’t matter, that you should not kill people, not mistreat parents or children, not steal, and not want someone else’s stuff (including wife) rather than going out and getting your own?</p></blockquote>
<p>Good point.  The Bible says nothing about shredders or psychology experiments.  So, everything&#8217;s cool.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/25/thou-shalt-not-cheat/feed/</wfw:commentRss>
			<slash:comments>17</slash:comments>
		
		
			</item>
		<item>
		<title>The fractal nature of scientific revolutions</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/24/the-fractal-nature-of-scientific-revolutions-3/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/24/the-fractal-nature-of-scientific-revolutions-3/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Fri, 24 Jan 2025 14:27:40 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51591</guid>

					<description><![CDATA[I was contacted by a new website, Daily 27&#8211;they asked me to write a short article of 300-500 words, &#8220;aimed at the general public&#8221; and related to &#8220;statistics and social sciences.&#8221; Considering what I&#8217;d written of this length that would &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/24/the-fractal-nature-of-scientific-revolutions-3/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>I was contacted by a new website, <a href="https://daily27.info">Daily 27</a>&#8211;they asked me to write a short article of 300-500 words, &#8220;aimed at the general public&#8221; and related to &#8220;statistics and social sciences.&#8221;  Considering what I&#8217;d written of this length that would be of general interest, I thought of the fractal nature of scientific revolutions, an idea that came up in <a href="http://stat.columbia.edu/~gelman/research/published/philosophy_chapter.pdf">this article with Cosma Shalizi</a>, Philosophy and the practice of Bayesian statistics in the social sciences, published in 2011 in the Oxford Handbook of the Philosophy of the Social Sciences.</p>
<p>The <a href="https://daily27.info/2025/01/23/the-fractal-nature-of-scientific-revolutions/">Daily 27 version is here</a>.</p>
<p>Oddly enough, they put it in the &#8220;Society and Gender&#8221; category.  I guess it makes sense&#8211;I&#8217;m a dude, and that&#8217;s a gender, so, sure, why not?</p>
<p><strong>P.S.</strong> I did some rooting around on the site and found <a href="https://daily27.info/2024/12/27/self-control-and-willpower-a-key-to-success-in-work-and-social-life/">an article</a> by the notorious Roy Baumeister.  <a href="https://statmodeling.stat.columbia.edu/2016/06/23/it-comes-down-to-reality-and-its-fine-with-me-cause-ive-let-it-slide/">Oh no</a> (see also <a href="https://statmodeling.stat.columbia.edu/2020/05/31/association-for-psychological-science-lays-down-the-law-criminal-justice-reform-and-unrestricted-sociosexuality-for-us-but-not-for-you/">here</a> and <a href="https://statmodeling.stat.columbia.edu/2020/08/03/kinda-like-reeses-pieces-if-you-dont-like-chocolate-and-you-dont-like-peanut-butter/">here</a>).  Oh well.  I&#8217;m not too proud to publish somewhere that also publishes Baumeister.  I also occasionally write for PNAS.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/24/the-fractal-nature-of-scientific-revolutions-3/feed/</wfw:commentRss>
			<slash:comments>5</slash:comments>
		
		
			</item>
		<item>
		<title>Science and the malleability of the self</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/23/science-and-the-malleability-of-the-self/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/23/science-and-the-malleability-of-the-self/#comments</comments>
		
		<dc:creator><![CDATA[Jessica Hullman]]></dc:creator>
		<pubDate>Thu, 23 Jan 2025 18:00:20 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Miscellaneous Statistics]]></category>
		<category><![CDATA[Sociology]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51589</guid>

					<description><![CDATA[&#8220;We still don’t really have any sure answers. We have made so much progress in some spheres, and yet so very little in others. We ought to feel the sting of our own ignorance; beset and bested on all sides &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/23/science-and-the-malleability-of-the-self/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<blockquote><p><span style="font-weight: 400">&#8220;We still don’t really have any sure answers. We have made so much progress in some spheres, and yet so very little in others. We ought to feel the sting of our own ignorance; beset and bested on all sides by the things we do not know. Why then do so many academics — and applied, business, whatever, “sciency” people — want to act as a colossus, act as if astride the world?&#8221;</span></p></blockquote>
<p><span style="font-weight: 400">This is Jessica. Rachael Meager wrote a </span><a href="https://rottenandgood.substack.com/p/taking-our-chances"><span style="font-weight: 400">really great essay</span></a><span style="font-weight: 400"> last week that ranged over many topics close to my heart (and perhaps the heart of this blog, if it has one), from superficial appeals to rigor and objectivity, heedless acceptance of generalities, disbelief in chance, and a tolerance (or even desire for) breathless predictions with no real mechanism for disincentivizing overconfidence. </span></p>
<p><span style="font-weight: 400">Ultimately it resonated as a meditation on how deep our discomfort runs when it comes to engaging with uncertainty as we construct our little research empires. These are themes that have been on my mind frequently over the last few years. I’d always been interested in limitations of scientific methods but it wasn’t until around the time I got tenure that I felt like I really “dove into the sea of uncertainty and stuck my head in and started blowing bubbles”, to paraphrase a quote from Andrew that Meager also reminded me of. One of the most striking parts of all this was realizing the extent to which people find uncertainty and doubt threatening, even as I was being careful to keep my opinions mostly to myself. </span><span style="font-weight: 400">It’s almost as if questioning the premises or foundations of certain research threatens the very “personhood” of the researchers involved. </span></p>
<p><span style="font-weight: 400">And so, in reading Meager’s post, I found myself thinking about how “avoiding the sting of ignorance” can go hand-in-hand with treating research as a kind of self-affirmation, where scientific investigation gets confused with the unfolding of the scientist’s personal narrative. And counterproductive norms develop to protect the researcher’s right to assert (and preserve) their self through their research.</span></p>
<p><span style="font-weight: 400">Meager writes:</span></p>
<blockquote><p><span style="font-weight: 400">Early on in my work I had to contend with people who thought quantitative evidence aggregation was unnecessary or useless in economics; later on some of these same people thought, well, look at the Cochrane handbook, this problem has already been solved. That’s sort of like looking at a stack of bibles and thinking we know everything about god. Actually people wrote the bible, and a lot of those people were fools.</span></p></blockquote>
<p><span style="font-weight: 400">One of the things I find most frustrating when some specific paper or line of work gets criticized as overconfident or unjustified is the seeming unwillingness on the part of the authors to risk something, to stop the knee jerk reaction to find a rationalization and instead actually sit with the possibility that they failed to consider something important. Instead, the fact that their work fits well into some larger landscape of research, where others have asked similar questions using similar methods gets interpreted as a sort of insurance against failing. I remember as a grad student learning that one technique for rebutting critical reviews on a paper you submitted is to find cases where the venue had published other papers that used the same method. Because if they did it and it was acceptable, why should I be held to a higher standard? And so we proceed as if it’s taken for granted that no criticism could ever be so bad as to motivate rescinding the idea.</span></p>
<blockquote><p><span style="font-weight: 400">I think most people do not really believe in inference, for inference means things are hidden. They only believe their own eyes. If something happens they think it obliterates all alternatives. It never could have been a different way. I think that lots of people think the whole world is completely deterministic.</span></p></blockquote>
<p><span style="font-weight: 400">As a result of this, there is enormous pressure to just go along with things. To say, “Okay, it’s fine” when someone presents you with some half-baked result, even after you just explained why you think it is quite far from fine and you don’t see how there is anything that can be said from the results at all. This is not to say that there is never any room for recognizing that the criticism makes a good point–this certainly happens sometimes, and sometimes the critical papers even win awards. It’s more that there is little precedent for truly taking back an idea. Once it’s out there, it must be right. And so once published, even the very bad ideas enjoy a kind of halo that paints them as still somehow valid in some way, and the authors of those papers are still justified in selling them without any real acknowledgement of the limitations. </span></p>
<p><span style="font-weight: 400">Andrew’s analogizing of publication as a kind of truth mill gets at part of this, but I’m interested in how this arises in part as a result of seeing research as a kind of affirmation of the self. It’s as if we’ve rewritten the definition of scientific learning to be something that stems from ourselves rather than the constraints of the world or of the representational systems that bind us in our particular fields. “Science as a celebration of the self” gets written into how we treat publications and academic milestones like dissertations, even as we claim to be interested only in “objective” truth. Denying that the work is ready gets confused with an attack on one’s right to exist. </span></p>
<blockquote><p><span style="font-weight: 400">People claim to care a lot about rigour, and exactitude, and chance. But people claim to care about a lot of things.</span></p>
<p><span style="font-weight: 400">With chance, in practice, people usually disrespect it. A lot of the time when people say they care about something, they just like the idea of the thing. To respect and to accept something is real is to accept that at times, like all things, it brings challenges. But that is not the way that we normally treat chance.</span></p></blockquote>
<p><span style="font-weight: 400">I was reviewing a paper recently where I was the only reviewer who disagreed with the premise of the work. To me the work was emblematic of how low expectations in this particular field about what constitutes a research contribution have come to validate authors churning out derivative “synthesizing” papers that do some shallow reflecting on trends in some emerging body of literature that was itself ad-hoc and reactive. It was so unambitious, a shot at the “minimum publishable unit”, with no clear generalizable contribution. I tried to get myself to think pragmatically, that maybe this paper was saying something other people would benefit from hearing even if I considered it a waste of time, but I couldn’t find it. And still, among the reviewers, I was the one who appeared to be out of touch. As if the fact that the paper was reasonably well-written and reacting to what had come before meant it was off the table that I could question it in this way. </span></p>
<p><span style="font-weight: 400">This attitude–that those who question the premises or take the holes too seriously must be wrong–is canonized through jokes about R2, the curmudgeonly reviewer who must have some personal affliction if they are trying to shoot down our work. I mean, how can they be justified when the </span><i><span style="font-weight: 400">majority of reviewers</span></i><span style="font-weight: 400"> agreed we were doing fine? Of course we deserve to go to the conferences and have a good time. That’s what we’re entitled to when we submit. If we do happen to get rejected, well, that&#8217;s when we believe in randomness and chance. We must have just gotten unlucky. It can’t be that we were wrong.</span></p>
<p><span style="font-weight: 400">So on some level it’s not that we never acknowlege uncertainty, it’s that we are only willing to accept it when it&#8217;s the wind that blows in the direction of our unfolding personal narrative. The converse of believing that some very hard problem has been “solved” because someone else has worked on it is refusing to accept that there can be a solution or that some real progress has been made, because that would constrain the development of our research identity. I see this a lot in research related to human-computer interfaces, whether it’s designing interactive data visualization systems or interfaces to AI/ML models.  For example, I get easily annoyed when people talk about topics like “appropriate reliance” on a predictive model or “evaluating AI explanations,” where often they can provide no clear definition of what they even mean. Meanwhile, decision theory provides a basis for clarifying much of this, but when you present them with such a perspective the response is like, &#8220;Well, that is one opinion&#8221; and they go right back to letting their intuition drive. </span></p>
<p><span style="font-weight: 400">Or, you point out that a paper already exists that technically solves the problem they care about but they don’t want to acknowledge this, because now that they had the idea that a solution is needed, it must come from them. It’s like we’re very good at pattern matching on the problem, but unwilling to do it on solutions. Once you’ve decided that you want to work on a problem, then it can’t be solved (at least, not by anyone but you). </span></p>
<blockquote><p><span style="font-weight: 400">The problem is that trying to really learn something new is humbling, especially if you come at it from a position of apparent success. The ego gets built up so, so easily, when things are going well, unless you work hard for it not to. It is not just online writers and clout-chasing posters who mistake numbers or success for importance, or for virtue. It is also real for tenured academics</span></p></blockquote>
<p><span style="font-weight: 400">The ego really does get in the way, and yet, we seem to welcome this. In computer science, we basically train for it. Advancing as a researcher coincides with imposing assertions, rather than learning to listen and slow down and study all of the details before speaking up. We’ve created an academic culture where being confused or at a loss for words has no place. Instead we’re often content to accept vibes in the guise of academic rigor. It’s easier to pretend the ego with its various hopes and dreams really is enough to suss out truth than to try to actually find it. </span></p>
<p><span style="font-weight: 400">And so it’s not necessarily surprising that many trends in how we publish and make hiring and promotion decisions point away from doing deep, thoughtful work, and toward pushing out as many assertions as one can. The expectation that everyone be publishing tens of papers a year, even as a Ph.D. student (at least in certain areas of computer science) is hardly likely to foster deep engagement with uncertainty. Ben Recht has called for </span><a href="https://substack.com/home/post/p-139922548"><span style="font-weight: 400">less publishing</span></a><span style="font-weight: 400">; similar proposals (judge papers based on quality, look at impact) seem to arise every few years since I’ve been a grad student, but so far none of these have made so much as a dent in the hyper-productivity. From the perspective that we’ve created a system with built-in safeguards against any serious threats to the researcher’s sense of self, is there really an alternative? If going deep is too risky, then we’re left with the churn. </span></p>
<p><span style="font-weight: 400">The funny thing is, even in the circumstances where there is the highest premium on being in complete control, like interviews or job talks, often it’s still possible to glimpse that there is an honest human beneath all of the storytelling, who does have a sense of what they don’t know. It’s just that we’ve trained researchers to be extremely cautious about admitting that. Better to pretend that you fully understand it all then to acknowledge in front of others that there’s still some confusion. </span></p>
<p><span style="font-weight: 400">As Meager writes, “</span><i><span style="font-weight: 400">there can be a real tradeoff between shooting for (mortal, fleeting) glory and carving out the time and space for good and deep and satisfying work of any kind.</span></i><span style="font-weight: 400">”</span></p>
<p><span style="font-weight: 400">Ultimately, what one works on, and how much uncertainty one chooses to let in, is a personal decision. The hard part of criticizing researchers for using their work as a form of self-affirmation or self-preservation is that of course we should expect research trajectories to be individualized. Everyone progresses in their thinking at their own rate, in directions that depend on their personal attractions to different topics and insecurities they perceive they must overcome. We become accustomed, even entrenched, in our preferred technical languages or ways of seeing, such that it’s hard to completely blame someone if they refuse to acknowledge a solution in another language, or to relate to doubts raised by someone they see as outside their preferred tradition. Trying to remove the personal entirely from scientific inquiry is unrealistic. </span></p>
<p><span style="font-weight: 400">And yet, I guess part of the sentiment behind this post, which Meager’s excellent essay really got me thinking about, is that as much as “the system” might seem to prevent us from going deep and internalizing doubt, ultimately it’s a personal choice. You can either choose to be the kind of scientist who’s in it for the personal glory or you can let uncertainty in. You can let “the standards” dictate how many papers you write a year and how ambitious each is, or you can try to be honest with yourself about when you’ve learned something worth sharing. There may be no way of omitting the self from science, but there are certainly different ways of using it. Research can be about self-discovery, but if there&#8217;s no risk, how can there be growth?</span></p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/23/science-and-the-malleability-of-the-self/feed/</wfw:commentRss>
			<slash:comments>42</slash:comments>
		
		
			</item>
		<item>
		<title>When estimating a treatment effect with a cluster design, you need to include varying slopes, even if the fit gives warning messages.</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/23/slopes/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/23/slopes/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Thu, 23 Jan 2025 14:50:01 +0000</pubDate>
				<category><![CDATA[Bayesian Statistics]]></category>
		<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Miscellaneous Statistics]]></category>
		<category><![CDATA[Multilevel Modeling]]></category>
		<category><![CDATA[Statistical Computing]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51585</guid>

					<description><![CDATA[Nick Brown and I recently published this paper, How statistical challenges and misreadings of the literature combineto produce unreplicable science: An example from psychology, in the journal, Advances in Methods and Practices in Psychological Science, about a published article that &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/23/slopes/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Nick Brown and I recently published this paper, <a href="http://stat.columbia.edu/~gelman/research/published/healing3.pdf">How statistical challenges and misreadings of the literature combineto produce unreplicable science: An example from psychology</a>, in the journal, Advances in Methods and Practices in Psychological Science, about a published article that claimed to find that healing of bruises could be sped or slowed by manipulating people’s subjective sense of time. The <a href="https://www.nature.com/articles/s41598-023-50009-3">article under discussion</a> had major problems&#8211;including weak theory, a flawed data analysis, and multiple errors in summarizing the literature&#8211;to the extent that we did not find their claims to be supported by their evidence.</p>
<p>That sort of thing happens&#8211;<a href="https://statmodeling.stat.columbia.edu/2016/07/08/29495/">we call it</a> everyday, bread-and-butter pseudoscience. The point of the new paper by Nick and me was not so much to shoot down one particular claim of mind-body healing but rather to explore some general reasons for scientific error and overconfidence and to provide a template for future researchers to avoid these problems in the future.</p>
<p>We posted this the other day on the blog (under the heading, <a href="https://statmodeling.stat.columbia.edu/2025/01/17/7steps/">7 steps to junk science that can achieve worldly success</a>), and something came up in comments, something I hadn&#8217;t thought of before and I wanted to share with you.  The commenter, Jean-Paul, <a href="https://statmodeling.stat.columbia.edu/2025/01/17/7steps/#comment-2388194">pointed to</a> a misunderstanding that can arise when fitting multilevel models using some software.</p>
<p><strong>Need for multilevel modeling or some equivalent adjustment for clustering in data</strong></p>
<p>Here was the problem.  The article in question estimated treatment effects with a cluster design: they had three treatments applied to 33 research participants (in psychology jargon, &#8220;subjects&#8221;) with data provided by 25 raters.  The total number of measurements was 2425 (not quite 3*33*25 because of some missing data, which is not the point in this story).</p>
<p>When you want to estimate treatment effects from a clustered design, you need to fit a multilevel model or make some equivalent adjustment.  Otherwise your standard errors will be too small.  In this case you should include effects for participants (some have more prominent bruises than others in this study) and for raters (who will have systematic differences in how they characterize the severity of a bruise using a numerical rating).</p>
<p>The researchers knew to do this, so they fit a multilevel model.  Fortunately for me, they used the R program lmer, which I&#8217;m familiar with.  Here&#8217;s the output, which for clarity I show using the display function in the arm package:</p>
<pre>
lmer(formula = Healing ~ Condition + (1 | Subject) + (1 | ResponseId), 
    data = DFmodel, REML = FALSE)
            coef.est coef.se
(Intercept) 6.20     0.31   
Condition28 0.23     0.09   
Condition56 1.05     0.09   

Error terms:
 Groups     Name        Std.Dev.
 Subject    (Intercept) 1.07    
 ResponseId (Intercept) 1.22    
 Residual               1.87    
---
number of obs: 2425, groups: Subject, 33; ResponseId, 25
AIC = 10131.4, DIC = 10119.4
deviance = 10119.4 
</pre>
<p>So, fine.  As expected, participant and rater effects are large&#8211;they vary by more than the magnitudes of the estimated treatment effects!&#8211;so it&#8217;s a good thing we adjusted for them.  And the main effects are a stunning 2.4 and 11.1 standard errors away from zero, a big win!</p>
<p><strong>Need for varying slopes, not just varying intercepts</strong></p>
<p>But . . . wait a minute. The point of this analysis is not just to estimate average responses; it&#8217;s to estimate treatment effects.  And, for that, we need to allow the treatment effects to vary by participant and rater&#8211;in multilevel modeling terms, we should be allowing slopes as well as intercepts to vary. This is in accordance with the well-known general principle of the design and analysis of experiments that the error term for any comparison should be at the level of analysis of the comparison.</p>
<p>No problem, lmer can do that, and we do so!  here&#8217;s the result:</p>
<pre>
lmer(formula = Healing ~ Condition + (1 + Condition | Subject) + 
    (1 + Condition | ResponseId), data = DFmodel, REML = FALSE)
            coef.est coef.se
(Intercept) 6.18     0.39   
Condition28 0.25     0.36   
Condition56 1.09     0.37   

Error terms:
 Groups     Name        Std.Dev. Corr        
 Subject    (Intercept) 1.71                 
            Condition28 1.99     -0.71       
            Condition56 2.03     -0.72  0.65 
 ResponseId (Intercept) 1.24                 
            Condition28 0.07      1.00       
            Condition56 0.13     -1.00 -1.00 
 Residual               1.51                 
---
number of obs: 2425, groups: Subject, 33; ResponseId, 25
AIC = 9317.9, DIC = 9285.9
deviance = 9285.9 
</pre>
<p>The estimated average treatment effects have barely changed, but the standard errors are much bigger.  The fitted models shows that treatment effects vary a lot by participant, not so much by rater.</p>
<p>You might be happy because the t-statistic for one of the treatment comparisons is still 3 standard errors away from zero, but other problems remain, both with multiple comparisons and with the interpretation of the data, as we discuss in section 2.4 of <a href="http://stat.columbia.edu/~gelman/research/published/healing3.pdf">our paper</a>.</p>
<p>There&#8217;s one thing you might notice if you look carefully at the above output, which is that the estimated covariance matrix for the rater effects is degenerate:  the correlations are at 1 and -1.  This sort of thing happens a lot with multilevel models when data are noisy and the number of groups is small.  It&#8217;s an issue we discussed in <a href="http://stat.columbia.edu/~gelman/research/published/chung_cov_matrices.pdf">this paper</a> from 2014.</p>
<p>In practice, it&#8217;s no big deal here:  the degeneracy is a consequence of a very noisy estimate, which itself arises because the noise in the data is much larger than the signal, which itself arises because the variation in treatment effects across raters is indistinguishable from zero (look at those small estimated standard deviations of the varying slopes for rater).  But that&#8217;s all pretty subtle, not something that&#8217;s in any textbook, even <a href="https://stat.columbia.edu/~gelman/arm/">ours</a>!</p>
<p><strong>That scary warning on the computer</strong></p>
<p>Here&#8217;s the problem.  When you fit that varying slopes model, R displays a warning in scary red type:</p>
<p><img decoding="async" src="https://statmodeling.stat.columbia.edu/wp-content/uploads/2025/01/Screenshot-2025-01-23-at-09.14.39-1024x158.png" alt="" width="450" /></p>
<p>It&#8217;s possible that the researchers tried this varying-slope fit, saw the singularity, got scared, and retreated to the simpler varying-intercept model, which had the incidental benefit of giving them an estimated treatment effect that was 11 standard error from zero.</p>
<p>Just to be clear, I&#8217;m not saying this is a problem with R, or with lme4.  Indeed, if you type help(&#8220;isSingular&#8221;) in the R console, you&#8217;ll get some reasonable advice, none of which is to throw out all the varying slopes.  But the first option suggested there is to &#8220;avoid fitting overly complex models in the first place,&#8221; which could  be misunderstood by users.</p>
<p>If you fit the model using stan_lmer on its default settings, everything works fine:</p>
<pre>
 family:       gaussian [identity]
 formula:      Healing ~ Condition + (1 + Condition | Subject) + (1 + Condition | 
	   ResponseId)
 observations: 2425
------
            Median MAD_SD
(Intercept) 6.22   0.40  
Condition28 0.26   0.34  
Condition56 1.09   0.37  

Auxiliary parameter(s):
      Median MAD_SD
sigma 1.51   0.02  

Error terms:
 Groups     Name        Std.Dev. Corr       
 Subject    (Intercept) 1.731               
            Condition28 2.015    -0.63      
            Condition56 2.056    -0.66  0.57
 ResponseId (Intercept) 1.225               
            Condition28 0.158     0.33      
            Condition56 0.172    -0.51  0.04
 Residual               1.509               
Num. levels: Subject 33, ResponseId 25 
</pre>
<p>These estimates average over the posterior distribution, so nothing on the boundary, no problem.  And, no surprise, the estimates and standard errors of the treatment effects are basically unchanged.</p>
<p>That said, when you run stan_lmer, it gives warnings in red too!</p>
<p><img decoding="async" src="https://statmodeling.stat.columbia.edu/wp-content/uploads/2025/01/Screenshot-2025-01-23-at-09.34.33-1024x212.png" alt="" width="650" /></p>
<p>These warning messages are a big problem!  On one hand, yeah, you want to warn people; on the other hand, it would be just horrible if the warnings are so scary that they send users to use bad models that don&#8217;t happen to spit out warnings.</p>
<p>I also tried fitting the model using blmer, which I thought would work fine, but that produced some warning messages that scared even me:</p>
<p><img decoding="async" src="https://statmodeling.stat.columbia.edu/wp-content/uploads/2025/01/Screenshot-2025-01-23-at-09.38.09-1024x169.png" alt="" width="550" /></p>
<p>By default, blme uses a degeneracy-avoiding prior, so this sort of thing shouldn&#8217;t happen at all. We should figure out what&#8217;s going on here!</p>
<p><strong>Summary</strong></p>
<p>1.  If you are estimating treatment effects using clustered data, you should be fitting a multilevel model with varying intercepts and slopes (or do the equivalent adjustment in some other way, for example by boostrapping according to the cluster structure).  Varying intercepts is not enough.  If you do it wrong, your standard errors can be way off.  This is all consistent with the general principle to use all design information in the analysis.</p>
<p>2.  If you apply a more complicated method on the computer and it gives you a warning message, this does <em>not</em> mean that you should go back to a simpler method.  More complicated models can be harder to fit.  This is something you might have to deal with.  If it really bothers you, then be more careful in your data collection; you can design studies that can be analyzed more simply.  Or you can do some averaging of your data, which might lose some statistical efficiency but will allow you to use simpler statistical methods (for example, see section 2.3, &#8220;Simple paired-comparisons analysis,&#8221; of <a href="http://stat.columbia.edu/~gelman/research/published/healing3.pdf">our paper</a>).</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/23/slopes/feed/</wfw:commentRss>
			<slash:comments>18</slash:comments>
		
		
			</item>
		<item>
		<title>Dan Ariely:  &#8220;Why Louisiana’s Ten Commandments law is a broken moral compass&#8221;</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/22/ariely-why-louisianas-ten-commandments-law-is-a-broken-moral-compass/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/22/ariely-why-louisianas-ten-commandments-law-is-a-broken-moral-compass/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Wed, 22 Jan 2025 14:37:17 +0000</pubDate>
				<category><![CDATA[Decision Analysis]]></category>
		<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51044</guid>

					<description><![CDATA[You think I&#8217;m kidding with the above title, but I&#8217;m not. Key quote from the noted social psychologist / business-school professor: The assumption that the Ten Commandments can serve as a universal moral code is increasingly out of touch with &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/22/ariely-why-louisianas-ten-commandments-law-is-a-broken-moral-compass/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>You think I&#8217;m kidding with the above title, <a href="https://thehill.com/opinion/education/4768231-louisiana-ten-commandments-law/">but I&#8217;m not</a>.  Key quote from the noted social psychologist / business-school professor:</p>
<blockquote><p>The assumption that the Ten Commandments can serve as a universal moral code is increasingly out of touch with contemporary American society.</p></blockquote>
<p>I didn&#8217;t check out all the links in this article, but my guess is that Ariely&#8217;s problem with the 10 Commandments is that God <a href="https://www.newyorker.com/magazine/2023/10/09/they-studied-dishonesty-was-their-work-a-lie">signed the tablets</a> at the bottom, not at the top.</p>
<p>Or maybe he didn&#8217;t like that commandment, &#8220;Thou shalt not make unto thee any graven image,&#8221; which, when translated into the language of modern science, might imply, &#8220;Thou shalt not fabricate data.&#8221;  Those damn commandments, always getting in the way!</p>
<p><strong>P.S.</strong>  You might say, why write about this?  Why not just decorously look away.  Well, remember what they say about <a href="https://statmodeling.stat.columbia.edu/2019/05/02/one-should-always-beat-a-dead-horse-because-the-horse-is-never-really-dead/">dead horses</a>.  The Hill is a reputable publication, and they&#8217;re using their space to promote questionable work.  I think this kind of thing is bad for science as a whole.  It&#8217;s just particularly ridiculous when they invite someone with major ethical concerns about his research to be lecturing us on ethics.  I do think mockery is appropriate here.  It&#8217;s nothing personal.  I just think of all the people in social science, working so hard and not pushing the ethical boundaries, working openly and honestly and with a sincere willingness to learn . . . and then this is what gets rewarded with publicity.  I laugh because otherwise I&#8217;d scream.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/22/ariely-why-louisianas-ten-commandments-law-is-a-broken-moral-compass/feed/</wfw:commentRss>
			<slash:comments>24</slash:comments>
		
		
			</item>
		<item>
		<title>&#8220;Interrogating Ethnography&#8221;:  The Alice Goffman story</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/21/interrogating-ethnicity-the-alice-goffman-story/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/21/interrogating-ethnicity-the-alice-goffman-story/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Tue, 21 Jan 2025 14:33:07 +0000</pubDate>
				<category><![CDATA[Sociology]]></category>
		<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=50595</guid>

					<description><![CDATA[I came across this book from 2018, Interrogating Ethnography: Why Evidence Matters, by law professor Steven Lubet. It&#8217;s a crisp (137 pages) and fascinating discussion of the role of evidence in qualitative social science, and I think it should be &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/21/interrogating-ethnicity-the-alice-goffman-story/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>I came across this book from 2018, Interrogating Ethnography:  Why Evidence Matters, by law professor Steven Lubet.  It&#8217;s a crisp (137 pages) and fascinating discussion of the role of evidence in qualitative social science, and I think it should be of interest to many of you, as it parallels so many discussions we&#8217;ve had over the years regarding the role of evidence in quantitative research.</p>
<p>Sometimes I&#8217;ve had negative reactions to writings by law professors on social science, but in this case Lubet&#8217;s expertise is relevant, as so many legal cases turn on evidence.</p>
<p>Lubet discusses several examples, focusing on sociologist Alice Goffman&#8217;s controversial 2015 book On the Run.  As we <a href="https://statmodeling.stat.columbia.edu/2016/01/20/rogue-sociologist-cant-stop-roguin/">discussed a few years ago</a>, it&#8217;s a problem of trust.  Goffman offers no documentation for her extraordinary claims and thus must rely on her readers and colleagues to trust her statements and treat them as fact.  In this case, trust is brittle, and once the trust is gone, not much remains.</p>
<p>One reason Lubet&#8217;s book is interesting is that he gets into the details and presents things very carefully.  Just for example, from page 131:</p>
<blockquote><p>It is unfortunate that ethnographers have so seldom essayed revisits to others&#8217; research sites.  Despite the obvious difficulties, there are cases in which the impediments can be readily overcome.  It would not take long for an ethnographer to interview personnel at the hospitals in West Philadelphia where Alice Goffman claims to have seen police cordons at the entrances.  Moreover, there are only six hospitals in Philadelphia with maternity services, so it would be possible, even now, to fact-check Goffman&#8217;s story of having observed the arrests of three new fathers on the same ward in a single evening.</p></blockquote>
<p>I&#8217;m guessing that this maternity ward falls <a href="https://statmodeling.stat.columbia.edu/2022/01/06/shreddergate-and-an-idea-for-a-museum-of-scholarly-misconduct/">into the same category as</a> Marc Hauser’s monkey tapes, Brian Wansink’s bottomless soup bowl and his 80-pound rock, Diederik Stapel’s survey forms, Mary Rosh’s survey forms, Michael Bellesiles’s probate inventories, Matthew Walker&#8217;s National Geographic video, the Surgisphere dataset, and Dan Ariely&#8217;s paper shredder.  But all things are possible.</p>
<p>The other thing notable about Lubet&#8217;s book is its even tone.  Some of the stories in the book are funny, others are kinda shocking, and Lubet manages to convey all this without himself ever expressing amusement or outrage.  There&#8217;s nothing wrong with expressing amusement or outrage&#8212;I do it all the time!&#8212;; it&#8217;s just impressive to me how he wrote this entire book with a straight face.  I recommend it.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/21/interrogating-ethnicity-the-alice-goffman-story/feed/</wfw:commentRss>
			<slash:comments>32</slash:comments>
		
		
			</item>
		<item>
		<title>The &#8220;delay-the-reckoning heuristic&#8221; in pro football?</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/20/the-delay-the-reckoning-heuristic-in-pro-football/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/20/the-delay-the-reckoning-heuristic-in-pro-football/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Mon, 20 Jan 2025 14:03:32 +0000</pubDate>
				<category><![CDATA[Decision Analysis]]></category>
		<category><![CDATA[Sports]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51065</guid>

					<description><![CDATA[Paul Campos tells this story of an NFL coach making a bad decision: Denver kicked a field goal with 1:54 to go to make the score 13-6 Pittsburgh. Denver had one time out remaining. At this point [Denver coach] Payton &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/20/the-delay-the-reckoning-heuristic-in-pro-football/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Paul Campos tells <a href="https://www.lawyersgunsmoneyblog.com/2024/09/statistical-estimates-on-the-fly-the-sean-payton-saga">this story</a> of an NFL coach making a bad decision:</p>
<blockquote><p>Denver kicked a field goal with 1:54 to go to make the score 13-6 Pittsburgh. Denver had one time out remaining. At this point [Denver coach] Payton elected to kick the ball off through the end zone rather than try an onside kick. If you do the math, this meant that the reasonable best case scenario for Denver was that they would stop Pittsburgh from getting a first down after three running plays, and get the ball back after a punt deep in their own territory with about ten seconds left and no time outs. This is in fact what happened. Since the odds of scoring a game-tying TD in this situation are almost zero, the question that naturally arose after the game is why Payton didn’t try an onside kick. . . .</p>
<p>Now the dumbest part of all this is that the downside of a failed onside kick in this situation is trivial. If the kick fails, Pittsburgh gets the ball around the 50 rather than on their own 30 after the kick through the end zone. If Pittsburgh makes a first down in either situation the game is over so that’s irrelevant. But what’s the downside of Pittsburgh punting from the Denver 44 instead of, as they did, from the Pittsburgh 36? This is a 20-yard difference, but, because the end zone serves as a constraint on punters since a punt that goes into the end zone comes out to the 20, the real difference in field position as a practical matter is probably more like ten yards, with the likely outcome being Denver getting the ball on its own 20 rather than its own ten. So Payton passed up a chance to get the ball back at midfield with nearly two minutes and a time out left — a very manageable situation if you need a TD — for a realistic best case scenario that would have required something tantamount to a miracle to result in a TD.</p></blockquote>
<p>Campos&#8217;s framing of the decision, a framing that seems reasonable to me, is that Denver, down by 7 points with 1:54 to go, had two options with their kickoff:</p>
<p>1.  Deep kick, then try to keep Pittsburgh from getting a first down, then try to score a touchdown with the one or two plays remaining after the punt recovery.</p>
<p>2.  Onside kick, then in the unlikely event that Denver recovers (approximately 6% chance, according to <a href="https://operations.nfl.com/gameday/analytics/stats-articles/where-have-all-the-points-gone/">this source</a>), they have more than a minute and a half to try to reach the end zone.</p>
<p>Campos argues that option 2 is better because, even conditional on stopping the first down and getting the ball back, the probability of scoring a touchdown in one or two plays, starting from deep in your own territory, is so much lower than the probability of scoring a touchdown from closer to midfield with a minute and a half to go.  If Denver has a 60% chance of stopping Pittsburgh from getting the first down, then this would imply that Campos thinks that Denver would be much less than 1/10th as likely to get the touchdown in one or two plays than with a minute and a half and one time out.  I don&#8217;t know these probabilities, but I assume a pro football coach would have an assistant who&#8217;d be able to access these numbers instantly.  As Campos says, &#8220;all of this happened immediately after the two minute warning, so he and his staff didn’t have to make snap decision: they had three minutes of beer and ED commercials to figure out this statistical puzzle.&#8221;</p>
<p>Assuming that (a) Campos is correct about the probabilities, and (b) that the decision isn&#8217;t even close, the question then arises, why did the coach make the wrong decision (which, again, we&#8217;re assuming was wrong prospectively, not just retrospectively)?</p>
<p>Campos offers three explanations:</p>
<blockquote><p>(1) When choosing between a course of action that creates a very slim chance of winning — onside kicks are rarely recovered — and one that creates a much slimmer chance, coaches tend to treat the decision in an irrational way, because what’s the difference between a one in 50 chance of winning and a one in 200 . . . when the baseline is low enough it’s like it doesn’t even matter. Now this is very likely true in this one particular instance, but it’s very much NOT true in the long run, when making many similar decisions over time.</p>
<p>(2) Coaches have a strong and strongly irrational preference for delaying the arrival of certain defeat for as long as possible, even at the cost of greatly reducing the odds of actually winning. . . . [This fits] Payton’s mentality throughout this game, including the decision to kick a field goal while down 13-0 with ten minutes to go, and even more so the decision to punt on fourth and eight from the Denver 33 while down by ten with seven and half minutes to go. Such decisions are both more likely to cause the moment of certain defeat to arrive later than it would arrive otherwise, and seriously suboptimal in terms of increasing the chances of actually winning the game.</p>
<p>(3) A third factor in such decisions is that coaches would prefer defeat while pursuing the conventional course of action to defeat while doing something unconventional, since the latter makes them prone to heavier criticism, even if the criticism is wrong. . . .</p></blockquote>
<p>Let&#8217;s get that third factor out of the way first, as it often comes up in this sort of discussions of coaching decisions.  We <a href="https://statmodeling.stat.columbia.edu/2022/11/20/going-for-it-on-4th-down-whats-striking-is-not-so-much-that-we-were-wrong-but-that-we-had-so-little-imagination-that-we-didnt-even-consider-the-possibility-that-we-might-be-wrong-i-wonder-wh/">talked about it</a> a few years ago in the context of fourth-down decisions in the NFL (with <a href="https://statmodeling.stat.columbia.edu/2022/11/22/4th-down-update-and-my-own-cognitive-illusion/">followup here</a>).  The short answer is yeah, most coaches have a motivation to be conservative, as it looks worse if you do something bold and it fails, but in this case the decision seems so clear that I don&#8217;t think the onside kick is particularly controversial.</p>
<p>As for the first factor . . . sure, I guess the point is that when a decision seems unimportant (in this case, raising the probability of tying or winning from near-zero to higher but still very low), then there is less motivation to make the rational decision.  But, I don&#8217;t really buy this.  Think about it the other way.  From the coach&#8217;s perspective, if the probability of winning is tiny, then the game is already basically lost, so at that point the team is playing with the house&#8217;s money.  So why not roll the dice?  From a psychological standpoint, this seems fundamentally different from if you&#8217;re almost certain to win and you make a seemingly-more conservative play even if it slightly increases your probability of losing, because if you&#8217;re gonna lose, you don&#8217;t want it to happen from what would be considered a weird play.</p>
<p>So then it comes down to Campos&#8217;s second argument, which interests me because it seems related to other decision-analysis paradoxes.  But, like other ideas in the always-confusing heuristics-and-biases literature, it introduces its own challenges.</p>
<p><strong>The delay-the-reckoning heuristic</strong></p>
<p>I&#8217;m gonna label this idea identified by Campos, that &#8220;Coaches have a strong and strongly irrational preference for delaying the arrival of certain defeat for as long as possible, even at the cost of greatly reducing the odds of actually winning,&#8221; as the <em>delay-the-reckoning heuristic</em>.</p>
<p>The general scenario is that you are at a fork in the decision tree, where one branch will give you decisive, or near-decisive, information right away, while the other branch will take you down further steps until the uncertainty is resolved.</p>
<p>Which fork will you take?</p>
<p>Just speaking in general terms, I feel like it could go either way.  Let me break the scenario into two sub-scenarios:  potential good news (as in the football example when you&#8217;re behind but there&#8217;s an outside chance you could get lucky and win) or potential bad news (if you were leading in the football game and there&#8217;s an outside chance you could get unlucky and lose).  Or you could think of medical examples:  you have a serious disease but there&#8217;s a potential miracle drug with a small chance of working, or you&#8217;re healthy and you&#8217;re going to take a blood test that might reveal you have an incurable cancer.</p>
<p>In the potential-good-news scenario, I agree with Campos that it somehow seems more natural (whatever that means) to delay the reckoning.  The idea is that you&#8217;re pretty sure it&#8217;s gonna be bad news, so you&#8217;d like to prolong the period of hope for as long as possible.  From a pure decision-analysis standpoint, it&#8217;s always better to get information sooner, as it can inform later steps in the decision problem, but from an emotional standpoint, I understand the appeal of keeping hope alive for as long as possible.</p>
<p>On the other hand, what about the &#8220;Give it to me straight, Doc?&#8221; attitude?  If you&#8217;re in bad shape, maybe you want to just know already and move on.  So I&#8217;m not sure.</p>
<p>Also, I kind of understand the emotional logic to delaying the reckoning . . . but this is all happening within two minutes of a football game!  Is it really worth lowering your win probability just to gain an additional minute and a half of hope?  That doesn&#8217;t seem quite right.  I feel like there&#8217;s something more going on here.</p>
<p>What about the potential-bad-news scenario?  There I feel like it&#8217;s natural to want the information as soon as possible so as to rule out the unlikely bad outcome.  Or maybe not.  I feel like I&#8217;m working in the grand tradition of judgment and decision making research, which is to theorize based on personal impressions of hypothetical scenarios.</p>
<p>I sent the above to Dan Goldstein, an expert in judgment and decision making, and he pointed us to the book, <a href="https://direct.mit.edu/books/edited-volume/5010/Deliberate-IgnoranceChoosing-Not-to-Know">Deliberate Ignorance: Choosing Not to Know</a>, edited by Ralph Hertwig and Christoph Engel.  So maybe there&#8217;s something there that&#8217;s relevant to our discussion.</p>
<p>The delay-the-reckoning heuristic interests me for its own sake and also for its connection to other time-related decision analysis issues such as the base-rate fallacy and its opposite, the <a href="https://statmodeling.stat.columbia.edu/2018/05/11/slow-to-update/">slow-to-update problem</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/20/the-delay-the-reckoning-heuristic-in-pro-football/feed/</wfw:commentRss>
			<slash:comments>21</slash:comments>
		
		
			</item>
		<item>
		<title>Problems caused by grade inflation</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/19/problems-caused-by-grade-inflation/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/19/problems-caused-by-grade-inflation/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Sun, 19 Jan 2025 14:52:36 +0000</pubDate>
				<category><![CDATA[Decision Analysis]]></category>
		<category><![CDATA[Economics]]></category>
		<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Teaching]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51040</guid>

					<description><![CDATA[Columbia math lecturer Peter Woit writes: There has been significant grade inflation over the years, so having a transcript with a string of As isn’t worth what it once was. This is not good for the unusually talented, who now &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/19/problems-caused-by-grade-inflation/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Columbia math lecturer Peter Woit <a href="https://www.math.columbia.edu/~woit/wordpress/?p=12069&#038;cpage=2#comment-238005">writes</a>:</p>
<blockquote><p>There has been significant grade inflation over the years, so having a transcript with a string of As isn’t worth what it once was. This is not good for the unusually talented, who now need to find other ways to distinguish themselves.</p></blockquote>
<p>That&#8217;s a good point!  I&#8217;ve typically thought of grade inflation in isolation (as in my post asking <a href="https://statmodeling.stat.columbia.edu/2011/07/27/12383/">why weren’t the instructors all giving all A’s already?</a>) with the problem being that inflated grades provide less information to future employers.</p>
<p>Woit&#8217;s point is related but goes further.  Now that A&#8217;s are given out like candy corn in the world&#8217;s worst Halloween party, they don&#8217;t provide much signal, first because, as Woit says, non-unusually-talented students can also get strings of A&#8217;s on their transcripts, and also because if you&#8217;re competing on grades, the occasional slip can be so costly.  Either way, ambitious students have to distinguish themselves in other ways&#8212;for example, by publishing articles in journals and conferences.  This propagation of &#8220;publish or perish&#8221; <a href="https://www.chronicle.com/article/teens-are-doing-ai-research-now-is-that-a-good-thing">down to the high school level</a> just exacerbates the explosion of publications&#8212;apparently, zillions of medical students are kinda required to publish research too, and if publication is a requirement, then the quality is not gonna matter so much, and these papers just get stirred in with whatever remaining legitimate literature is being produced.</p>
<p>So, yeah, if we were to give out more B&#8217;s and C&#8217;s, maybe the world would be a better place.</p>
<p>I&#8217;m not planning to first, though.  As I wrote <a href="https://statmodeling.stat.columbia.edu/2011/07/27/12383/">a few years ago</a>, the real mystery to me is not, Why is there grade inflation?, but rather, Why is there any room left to inflate: why weren’t the instructors all giving all A’s already?</p>
<p>At that time, I recommended statistician Val Jonhson&#8217;s plan to &#8220;make post-hoc adjustments to assigned grades to account for differences in faculty grading policies&#8221;&#8212;basically, fit a multilevel item-response model to estimate students&#8217; latent abilities based on their grades.  As I wrote at the time:</p>
<blockquote><p>The beauty of <a href="https://projecteuclid.org/journals/statistical-science/volume-12/issue-4/An-alternative-to-traditional-GPA-for-evaluating-student-performance/10.1214/ss/1030037959.full">Val’s approach</a> is that it does three things:</p>
<p>1. By statically correcting for grading practices, Val’s method produces adjusted grades that are more informative measures of student ability.</p>
<p>2. Since students know their grades will be adjusted, they can choose and evaluate their classes based on what they expect to learn and how they expect to perform; they don’t have to worry about the extraneous factor of how easy the grading is.</p>
<p>3. Since instructors know the grades will be adjusted, they can assign grades for accuracy and not have to worry about the average grade. (They can still give all A’s but this will no longer be a benefit to the individual students after the course is over.)</p></blockquote>
<p>I still like Val&#8217;s idea, but at this point there may be too much grade inflation at some schools for it to work.  At some point there is so little signal left that you can&#8217;t recover the information you want.</p>
<p>OK, at this point you might say, sure, grades are B.S., whatever.  But that puts in the worse position of implicitly requiring students to have other qualifications.  At best, this sends students to interesting research projects and internships, but many times it just pushes them into trying to hop on projects to get credentials.  Rather than writing some crappy Neurips paper and then learning the tricks to get it accepted, I think they&#8217;d be better off taking interesting courses in college, working hard, doing well on exams, and writing good term papers.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/19/problems-caused-by-grade-inflation/feed/</wfw:commentRss>
			<slash:comments>53</slash:comments>
		
		
			</item>
		<item>
		<title>Where should we publish our paper, &#8220;Statistical graphics and comics: Parallel histories of visual storytelling&#8221;?</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/18/where-should-we-publish-our-paper-statistical-graphics-and-comics-parallel-histories-of-visual-storytelling/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/18/where-should-we-publish-our-paper-statistical-graphics-and-comics-parallel-histories-of-visual-storytelling/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Sat, 18 Jan 2025 14:21:48 +0000</pubDate>
				<category><![CDATA[Art]]></category>
		<category><![CDATA[Literature]]></category>
		<category><![CDATA[Statistical Graphics]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51573</guid>

					<description><![CDATA[Hey! Susan Kruglinski and I wrote this article I really like, Statistical graphics and comics: Parallel histories of visual storytelling: What do data visualization and cartoons have in common? One of these is used to communicate in science and journalism, &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/18/where-should-we-publish-our-paper-statistical-graphics-and-comics-parallel-histories-of-visual-storytelling/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Hey!  Susan Kruglinski and I wrote this article I really like, <a href="http://stat.columbia.edu/~gelman/research/unpublished/graphics_bd.pdf">Statistical graphics and comics: Parallel histories of visual storytelling</a>:</p>
<blockquote><p>What do data visualization and cartoons have in common? One of these is used to communicate in science and journalism, and the other appears in arts and entertainment, but both convey complex messages in economical, intuitive, and visually appealing ways. And both these graphic forms are relatively new, having made rapid progress only in the past few centuries, despite requiring little in the way of raw material to produce. We connect this history to a combination of abstraction and accessibility that is common to both these forms of visual expression: comic strips and scatterplots both now seem intuitive but represent the development of abstract conventions. We also discuss differences between these two methods of visual storytelling in their goals and in how they are experienced by the reader.</p></blockquote>
<p>Read <a href="http://stat.columbia.edu/~gelman/research/unpublished/graphics_bd.pdf">the whole thing</a>.  It has a message I think is important.</p>
<p>But my message to you is:  Where should we publish this article?  We sent it to the journal American Statistician, which didn&#8217;t seem quite right; in any case they agreed with that assessment and told us it would be better to publish somewhere else.  But we&#8217;re not sure where.</p>
<p>There&#8217;s no need for the paper to appear in a statistics journal, or in a &#8220;journal&#8221; at all&#8211;it&#8217;s not like we&#8217;re getting &#8220;publish or perish&#8221; credit for it! Lots of non-statistician &#8220;civilians&#8221; are interested in dataviz and comics, and I&#8217;d like to reach some audience beyond whoever&#8217;s reading this post right now.</p>
<p>If you have any thoughts on where to publish this article&#8211;or, of course, any thoughts on the substance of the article itself&#8211;you can just let us know right here in the comments section.  Otherwise, just enjoy the article.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/18/where-should-we-publish-our-paper-statistical-graphics-and-comics-parallel-histories-of-visual-storytelling/feed/</wfw:commentRss>
			<slash:comments>22</slash:comments>
		
		
			</item>
		<item>
		<title>How far can exchangeability get us toward agreeing on individual probability?</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/17/how-far-can-exchangeability-get-us-toward-agreeing-on-individual-probability/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/17/how-far-can-exchangeability-get-us-toward-agreeing-on-individual-probability/#comments</comments>
		
		<dc:creator><![CDATA[Jessica Hullman]]></dc:creator>
		<pubDate>Fri, 17 Jan 2025 17:39:56 +0000</pubDate>
				<category><![CDATA[Bayesian Statistics]]></category>
		<category><![CDATA[Causal Inference]]></category>
		<category><![CDATA[Decision Analysis]]></category>
		<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Miscellaneous Statistics]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51571</guid>

					<description><![CDATA[This is Jessica. What’s the common assumption behind the following?  Partial pooling of information over groups in hierarchical Bayesian models  In causal inference of treatment effects, saying that the outcome you would get if you were treated (Y^a) shouldn’t change &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/17/how-far-can-exchangeability-get-us-toward-agreeing-on-individual-probability/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><span style="font-weight: 400">This is Jessica. What’s the common assumption behind the following? </span></p>
<ul>
<li style="list-style-type: none">
<ul>
<li style="font-weight: 400"><span style="font-weight: 400">Partial pooling of information over groups in hierarchical Bayesian models </span></li>
<li style="font-weight: 400"><span style="font-weight: 400">In causal inference of treatment effects, saying that </span><span style="font-weight: 400">the </span><span style="font-weight: 400">outcome you would get if you were treated (Y^a) shouldn’t change depending on whether you are assigned the treatment (A) or not</span></li>
<li style="font-weight: 400"><span style="font-weight: 400">Acting as if we believe a probability is the “objective chance” of an event even if we prefer to see probability as an assignment of betting odds or degrees of belief to an event</span></li>
</ul>
</li>
</ul>
<p><span style="font-weight: 400">The question is rhetorical, because the answer is in the post title. These are all examples where statistical exchangeability is important. Exchangeability says the joint distribution of a set of random variables is unaffected by the order in which they are observed. </span></p>
<p><span style="font-weight: 400">Exchangeability has broad implications. Lately I’ve been thinking about it as it comes up at the ML/stats intersection, where it’s critical to various methods: achieving coverage in </span><a href="https://arxiv.org/abs/2107.07511"><span style="font-weight: 400">conformal prediction</span></a><span style="font-weight: 400">, using counterfactuals in </span><a href="https://dl.acm.org/doi/pdf/10.1145/3351095.3372851"><span style="font-weight: 400">analyzing algorithmic fairness</span></a><span style="font-weight: 400">, </span><a href="https://arxiv.org/pdf/2405.18836"><span style="font-weight: 400">identifying independent causal mechanisms</span></a><span style="font-weight: 400"> in observational data, etc. </span></p>
<p><span style="font-weight: 400">This week it came up in the course I’m teaching on </span><a href="https://statmodeling.stat.columbia.edu/2024/12/06/new-course-prediction-for-individualized-decision-making/"><span style="font-weight: 400">prediction for decision-making</span></a><span style="font-weight: 400">. A student asked whether exchangeability was of interest because often people aren’t comfortable assuming data is IID. I could see how this might seem like the case given how application-oriented papers (like on conformal prediction) sometimes talk about the exchangeabilty requirement as an advantage over the usual assumption of IID data. But this misses the deeper significance, which is that it provides a kind of practical consensus between different statistical philosophies. This consensus, and the ways in which it’s ultimately limited, is the topic of this post.</span></p>
<p><b>Interpreting the probability of an individual event</b></p>
<p><span style="font-weight: 400">One of the papers I’d assigned was Dawid’s “On Individual Risk,” which, as you might expect, talks about what it means to assign probability to a single event. Dawid distinguishes “groupist” interpretations of probability that depend on identifying some set of events, like the frequentist definition of probability as the limiting frequency over hypothetical replications of the event, from individualist interpretations, like a “personal probability” reflecting the beliefs of some expert about some specific event conditioned on some prior experience. For the purposes of this discussion, we can put Bayesians (subjective, objective, and pragmatic, as Bob describes them </span><a href="https://statmodeling.stat.columbia.edu/2024/07/10/three-cultures-bayes-subjective-objective-pragmatic/"><span style="font-weight: 400">here</span></a><span style="font-weight: 400">) in the latter personalist-individualist category. </span></p>
<p><span style="font-weight: 400">On the surface, the frequentist treatment of probability as an “objective” quantity appears incompatible with the individualist notion of probability as a descriptor of a particular event from the perspective of the particular observer (or expert) ascribing beliefs. If you have a frequentist and a personalist thinking about the next toss of a coin, for example, you would expect the probability the personalist assigns to depend on their joint distribution over possible sequences of outcomes, while the frequentist would be content to know the limiting probability. But de Finetti’s theorem shows that if one believes a sequence of events to be exchangeable, then you can’t distinguish their beliefs about those random variables from conceiving of independent events with some underlying probability. Given a sequence of exchangeable Bernoulli random variables X1, X2, X3, … , you can think of a draw from their joint distribution as sampling </span><i><span style="font-weight: 400">p ~ mu</span></i><span style="font-weight: 400">, then drawing X1, X2, X3, … from </span><i><span style="font-weight: 400">Bernoulli(p)</span></i><span style="font-weight: 400"> (where mu is a distribution on [0,1]). </span><span style="font-weight: 400">So the frequentist and personalist can both agree, under exchangeability, that </span><i><span style="font-weight: 400">p</span></i><span style="font-weight: 400"> is meaningful for decision making. David Spiegalhalter recently published an </span><span style="font-weight: 400"><a href="https://www.nature.com/articles/d41586-024-04096-5">essay on interpreting probability</a> that he ended by commenting on how remarkable this pragmatic consensus is</span><span style="font-weight: 400">.</span></p>
<p><span style="font-weight: 400">But Dawid’s goal is to point out ways in which the apparent alignment is not as satisfactory as it may seem in resolving the philosophical chasm. It’s more like we’ve thrown a (somewhat flimsy) plank over it. Exchangeability may sometimes get us across by allowing the frequentist and personalist to coordinate in terms of actions, but we have to be careful how much weight we put on this.  </span></p>
<p><b>The reference set depends on the state of information</b></p>
<p><span style="font-weight: 400">One complication is that the personalist’s willingness to assume exchangeability depends on the information they have. Dawid uses the example of trying to </span><span style="font-weight: 400">predict the exam score of some particular student. </span><span style="font-weight: 400">If they have no additional information to distinguish the target student from the rest, </span><span style="font-weight: 400">the personalist might be content to be given an overall limiting relative frequency </span><i><span style="font-weight: 400">p </span></i><span style="font-weight: 400">of failure across a set of students. But as soon as they learn something that makes the individual student unique, </span><i><span style="font-weight: 400">p </span></i><span style="font-weight: 400">is no longer the appropriate reference for the individual student’s probability of passing the exam. </span></p>
<p><span style="font-weight: 400">As an aside, this doesn’t mean that exchangeability is only useful if we think of members of some exchangeable set as identical. There may still be practical benefits of learning from the other students in the context of a statistical model, for example. See, e.g., </span><span style="font-weight: 400">Andrew’s previous post on exchangeability as an </span><a href="https://statmodeling.stat.columbia.edu/2022/11/24/understanding-exchangeability-in-statistical-modeling-a-thanksgiving-themed-post/"><span style="font-weight: 400">assumption in hierarchical models</span></a><span style="font-weight: 400">, where he points out that assuming exchangeability doesn’t necessarily mean that you believe everything is indistinguishable, and if you have additional information distinguishing groups, you can incorporate that in your model as group-level predictors.</span></p>
<p><span style="font-weight: 400">But for the purposes of personalists and frequentists agreeing on a reference for the probability of a specific event, the dependence on information is not ideal. Can we avoid this by making the reference set more specific? What if we’re trying to predict a particular student’s score on a particular exam in a world where that particular student is allowed to attempt the same exam as many times as they’d like? Now that the reference group refers to the particular student and particular exam, would the personalist be content to accept the limiting frequency as the probability of passing the next attempt? </span></p>
<p><span style="font-weight: 400">The answer is, not necessarily. This imaginary world still can’t get us to the generality we’d need for exchangeability to truly reconcile a personalist and frequentist assessment of the probability. </span></p>
<p><b>Example where the limiting frequency is constructed over time</b></p>
<p><span style="font-weight: 400">Dawid illustrates this by introducing a complicating (but not at all unrealistic) assumption: that the student’s performance on their next try on the exam will be affected by their performance on the previous tries. Now we have a situation where the limiting frequency of passing on repeated attempts is constructed over time. </span></p>
<p><span style="font-weight: 400">As an analogy, consider drawing balls from an urn, where when we draw our first ball, there is 1 red ball and 1 green ball in it. Upon drawing a ball, we immediately return and add an additional ball of the same color. At each draw, each ball in the urn is equally likely of being drawn, and  the sequence of colors is exchangeable. </span></p>
<p><span style="font-weight: 400">Given that </span><i><span style="font-weight: 400">p </span></i><span style="font-weight: 400">is not known, which do you think the personalist would prefer to consider as the probability of a red ball on the first draw: the proportion of red balls currently in the urn, or the limiting frequency of drawing a red ball over the entire sequence? </span></p>
<p><span style="font-weight: 400">Turns out in this example, the distinction doesn’t actually matter: the personalist should just bet 0.5. So why is there still a problem in reconciling the personalist assessment with the limiting frequency?</span></p>
<p><span style="font-weight: 400">The answer is that we now have a situation where knowledge of the dynamic aspect of the process makes it seem contradictory for the personalist to trust the limiting frequency. If they know it’s constructed over time, then on what ground is the personalist supposed to assume the limiting frequency is the right reference for the probability on the first draw? This gets at the awkwardness of using behavior in the limit to think about individual predictions we might make.</span></p>
<p><b>Why this matters in the context of algorithmic decision-making</b></p>
<p><span style="font-weight: 400">This example is related to some of my prior posts on why calibration does not satisfy everyone as a means of ensuring good decisions. The broader point in the context of the course I’m teaching is that when we’re making risk predictions (and subsequent decisions) about people, such as in deciding who to grant a loan or whether to provide some medical treatment, there is inherent ambiguity in the target quantity. Often there are expectations that the decision-maker will do their best to consider the information about that particular person and make the best decision they can. What becomes important is not so much that we can guarantee our predictions behave well as a group (e.g., calibration) but that we understand how we’re limited by the information we have and what assumptions we’re making about the reference group in an individual case. </span></p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/17/how-far-can-exchangeability-get-us-toward-agreeing-on-individual-probability/feed/</wfw:commentRss>
			<slash:comments>14</slash:comments>
		
		
			</item>
		<item>
		<title>7 steps to junk science that can achieve worldly success</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/17/7steps/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/17/7steps/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Fri, 17 Jan 2025 14:22:15 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Sociology]]></category>
		<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51566</guid>

					<description><![CDATA[More than a decade after the earthquake that was the replication crisis (for some background, see my article with Simine Vazire, Why did it take so many decades for the behavioral sciences to develop a sense of crisis around methodology &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/17/7steps/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>More than a decade after the earthquake that was the replication crisis (for some background, see my article with Simine Vazire, <a href="http://stat.columbia.edu/~gelman/research/published/jmmss-3062-gelman.pdf">Why did it take so many decades for the behavioral sciences to develop a sense of crisis around methodology and replication?</a>), it is frustrating to see junk science still being published, promoted, and celebrated, even within psychology, the field that was at the epicenter of the crisis.</p>
<p><strong>The crisis continues</strong></p>
<p>An example that I learned about recently was an article out of Harvard, <a href="https://www.nature.com/articles/s41598-023-50009-3">Physical healing as a function of perceived time</a>, published in 2023 and subsequently promoted in the news media, that claimed to demonstrate that healing of bruises could be sped or slowed by manipulating people&#8217;s subjective sense of time.  All things are possible, and never say never, but, yeah, this paper offered no good evidence for its extraordinary claims.  It was standard-issue junk science: a grabby idea, a statistically significant p-value extracted from noisy data, and big claims.</p>
<p>Someone pointed me to this paper, and for some reason that I can no longer remember, Nick Brown and I decided to figure out exactly what went wrong with it.  We published our findings in this article, <a href="http://stat.columbia.edu/~gelman/research/published/healing3.pdf">How statistical challenges and misreadings of the literature combine to produce unreplicable science: An example from psychology</a>, which will appear in the journal Advances in Methods and Practices in Psychological Science.</p>
<p>In short, the published article was flawed in two important ways, first in its statistical analysis (see section 2.4 of our paper, where we write, &#8220;We are skeptical that this study reveals anything about the effect of perceived time on physical healing, for four reasons&#8221;) and second in its interpretation of its cited literature (see section 3 of our paper, where we write, &#8220;Here we discuss three different examples of this sort of misinterpretation of the literature cited in the paper under discussion&#8221;).</p>
<p>I don&#8217;t have any particular interest in purported mind-body healing, but Nick and I went to the trouble to shepherd our article through the publication process, with two goals in mind:<br />
&#8211; Providing an example of how we, as outsiders, could look carefully at a research article and its references and figure out what went wrong. This is important, because it&#8217;s pretty common to see papers that make outlandish claims but seem to be supported by data and the literature.<br />
&#8211; Exploring what exactly goes wrong&#8211;in this case, it was a mis-analysis of a complex data structure, researcher degrees of freedom in decisions of what to report, and multiple inaccurate summaries of the literature.</p>
<p><strong>What does it take for junk science to be successful?</strong></p>
<p>All this got me thinking about what it takes for researchers to put together a successful work of junk science in the modern era, which is the subject of today&#8217;s post.</p>
<p>Before going on, let me emphasize that I have no reason to suspect misconduct on the part of the authors of the paper in question. It&#8217;s a bad paper, and it&#8217;s bad science, but that happens given how people are trained, and given the track record of what gets published in leading journals (Psychological Science, PNAS), what gets rewarded in academia, and what gets publicity from NPR, Ted, Freakonomics, and the like.  As we&#8217;ve discussed many times, you can do bad science without being a bad person and without committing what would usually be called research misconduct. (I actually don&#8217;t think that bad data analysis and inaccurate description of the literature would usually be put in the &#8220;research misconduct&#8221; category.)</p>
<p>This is also why I&#8217;m not mentioning the authors&#8217; names here.  The names are no secret&#8211;just click on the above link and the paper is right there!&#8211;I&#8217;m just not including them in this post, so as to emphasize that I&#8217;m writing here about the process of bad science and its promotion; it&#8217;s not about these particular authors (or any particular authors).</p>
<p><strong>7 steps to junk science</strong></p>
<p>So here they are, 7 things that allow junk science to thrive:</p>
<p><strong>1.  Bad statistical analysis.</strong>  Statistics is hard; there are a lot of ways to make mistakes, and often these mistakes can lead to what appears to be strong evidence.</p>
<p><strong>2.  Researcher degrees of freedom.</strong>  Garden of forking paths.  As always, the problem is not with the forking paths&#8211;there really are a lot of ways to collect, code, and analyze data!&#8211;but rather with selection in what is reported.  As Simmons et al. (2011) unforgettably put it, &#8220;undisclosed flexibility in data collection and analysis allows presenting anything as significant.&#8221;  And, as Loken and I emphasized in our paper on forking paths, &#8220;undisclosed flexibility&#8221; could be undisclosed to the authors themselves:  the problem is with data-dependent analysis choices, even if the data at hand were analyzed only once.</p>
<p><strong>3.  Weak or open-ended substantive theory.</strong>  Theories such as evolutionary psychology, embodied cognition, and mind-body healing are vague enough to explain just about anything.  As Brown and I wrote in our above-linked article, &#8220;The authors refer to &#8216;mind–body unity&#8217; and &#8216;the importance of psychological factors in all aspects of health and wellbeing,&#8217; and we would not want to rule out the possibility of such an eﬀect, but no mechanisms are examined in this study, so the result seems at best speculative, even taking the data summaries at face value. During the half hour of the experimental conditions, the participants were performing various activities on the computer that could affect blood flow, and these activities were different in each condition . . . there are many alternative explanations for the results which we find<br />
just as scientifically plausible as the published claim.&#8221;</p>
<p><strong>4.  Inaccurate summaries of the literature.</strong>  This is a big deal, a huge deal, and something we don’t talk enough about.</p>
<p>It’s a lot to expect the journal editors and reviewers to check citations and literature reviews. <em>It&#8217;s your job as an author</em> to read and understand the work you&#8217;re citing before using those papers to make unsupported claims.  For example, don&#8217;t make the claim, “If a person who does not exercise weighed themselves, checked their blood pressure, took careful body measurements, wrote everything down, maintained their same diet and level of physical activity, and then repeated the same measures a month later, few would expect exercise-like improvements. But in a study involving hotel housekeepers, that is effectively what the researchers found,&#8221; if you&#8217;re citing a study that does not support this claim.</p>
<p><strong>5.  Institutional support.</strong>  Respectable journals are willing to publish articles that make outlandish claims based on weak evidence.  Respected universities give Ph.D.&#8217;s for such work.  Again, I&#8217;m not suggesting malfeasance on the part of the authors; they&#8217;re just playing by the rules that they&#8217;ve learned.</p>
<p><strong>6.  External promotion.</strong>  This work was featured in Freakonomics, Scientific American, and other podcasts and news outlets (see <a href="https://statmodeling.stat.columbia.edu/2024/10/19/carroll-langer-credulous-scientist-as-hero-reporting-from-a-podcaster-who-should-know-better/">here</a> and <a href="https://statmodeling.stat.columbia.edu/2024/10/28/freakonomics-does-it-again-not-in-a-good-way-jeez-these-guys-are-credulous/">here</a>).  This external promotion has three malign effects:<br />
&#8211; Most directly, it spreads the (inaccurate) word about the bad research.<br />
&#8211; The publicity also provides an incentive for people to more sloppy work that can yield these sorts of strong claims from weak evidence.<br />
&#8211; Also, publicity for sloppy, bad science can <a href="https://statmodeling.stat.columbia.edu/2016/03/31/greshams-law-of-experimental-methods/">crowd out</a> publicity and reduce the incentives to do careful, good science.</p>
<p><strong>7.  Celebrity culture.</strong>  This is a combination of items 5 and 6 above:  many celebrity academic and media figures prop each other up. Some of it&#8217;s from converging interests, as when the Nudgelords <a href="https://statmodeling.stat.columbia.edu/2022/06/04/pizzagate-and-nudge-an-opportunity-lost/">presented</a> the work of Brian Wansink as &#8220;masterpieces,&#8221; but often I think it&#8217;s more just a sense that all these media-friendly scientists and podcasters and journalists feel that they&#8217;re part of some collective project of science promotion, and from that perspective it doesn&#8217;t really matter if the science is good or bad, as long as it&#8217;s science-like, by their standards.</p>
<p>Anyway, this continues to bug the hell out of me, which is why I keep chewing on it and writing about it from different angles.  I&#8217;m glad that Nick and I wrote that paper&#8211;it took some effort to track down all the details and express ourselves both clearly and carefully.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/17/7steps/feed/</wfw:commentRss>
			<slash:comments>22</slash:comments>
		
		
			</item>
		<item>
		<title>Why I like preregistration (and it&#8217;s not about p-hacking).  When done right, it unifies the substance of science with the scientific method.</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/16/why-i-like-preregistration-and-its-not-about-p-hacking-when-done-right-it-unifies-the-substance-of-science-with-the-scientific-method/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/16/why-i-like-preregistration-and-its-not-about-p-hacking-when-done-right-it-unifies-the-substance-of-science-with-the-scientific-method/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Thu, 16 Jan 2025 14:18:59 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Sociology]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51081</guid>

					<description><![CDATA[This came up in comments to Jessica&#8217;s recent post. I like preregistration. It’s not something I used to do, and I still don’t always do it. I’ve worked on hundreds of research projects, and only a few of them had &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/16/why-i-like-preregistration-and-its-not-about-p-hacking-when-done-right-it-unifies-the-substance-of-science-with-the-scientific-method/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>This came up <a href="https://statmodeling.stat.columbia.edu/2024/09/19/getting-a-pass-on-evaluating-ways-to-improve-science/#comment-2379860">in comments</a> to Jessica&#8217;s recent post.</p>
<p>I like preregistration. It’s not something I used to do, and I still don’t always do it. I’ve worked on hundreds of research projects, and only a few of them had had any preregistration at all.</p>
<p>That said, I think preregistration has value, and I’m doing it more and more.</p>
<p>The reason I like preregistration has nothing at all to do with hypothesis tests or p-values or p-hacking or questionable research practices or anything like that.</p>
<p>I like preregistration for two reasons.</p>
<p>1. For me, preregistration implies constructing a hypothetical world&#8211;not a “null hypothesis” of no effect, but a possible world corresponding to what I’m actually aiming to study&#8211;and then simulating fake data and proposing and trying out analysis methods on those simulated data. I find this sort of commitment&#8211;the effort of laying out a complete generative model for the process&#8211;to be helpful. Thinking about <a href="http://stat.columbia.edu/~gelman/research/unpublished/causal_quartets.pdf">effect sizes and their variation</a>, all sorts of things, also seeing if the proposed analysis can recover parameters of interest from the simulated data, which is what’s often called power analysis <a href="http://stat.columbia.edu/~gelman/research/published/retropower_final.pdf">although I prefer</a> the more general term “design analysis.”</p>
<p>2. When other people preregister, that can be useful because then we can see discrepancies between the original plan and what actually got reported. Two examples are <a href="https://statmodeling.stat.columbia.edu/2024/04/04/its-ariely-time-an-example-of-the-success-of-preregistration-in-revealing-problems-with-a-published-paper/">here</a> and <a href="https://statmodeling.stat.columbia.edu/2018/05/29/exposure-forking-paths-affects-support-publication/">here</a>&#8211;in both those cases, discrepancies between the preregistration and the final paper gave us doubts about the published claims. When these changes happen, it is not a moral failure on anyone’s part&#8211;we can learn from data!&#8211;it’s just relevant for understanding the theories being promulgated in these papers.</p>
<p>I agree that preregistration is not necessary for good science. I still think it can be a useful tool, both my own workflow in developing scientific hypotheses and gathering data to understand them, and in communication of workflow to others.</p>
<p>Preregistration has a valuable indirect function of making it more difficult to do bad science. It does not directly turn bad science into good science. That doesn’t make preregistration a bad idea&#8211;recently I’ve been preregistering studies and, more generally, <a href="https://statmodeling.stat.columbia.edu/2019/03/23/yes-i-really-really-really-like-fake-data-simulation-and-i-cant-stop-talking-about-it/">simulating data</a> before gathering any data&#8211;; we should just be aware that this sort of procedural step can only one small part of the story. Ultimately, science is about the substance of science, not just about the scientific method.</p>
<p>There’s something interesting here, though, that links the two perspectives. If you do things right, your preregistration will involve the substance of what you’re studying and will not merely be a procedural step, a form of paperwork that exists to validate the p-values that your study will produce. Rather, doing this preregistration will require simulating fake data, which in turn will require hypothesizing a full model of the underlying process.</p>
<p>I recognize that what I just described is not the usual thing that is meant by “preregistration,” which is more along the lines of: “We will perform this comparison and use a 2-sided test,” etc. But it could be! I think this is a useful connection.</p>
<p><strong>P.S.</strong>  As discussed in <a href="https://statmodeling.stat.columbia.edu/2025/01/16/why-i-like-preregistration-and-its-not-about-p-hacking-when-done-right-it-unifies-the-substance-of-science-with-the-scientific-method/#comment-2387800">comments</a>, a more precise term for what I&#8217;m recommending is <a href="https://statmodeling.stat.columbia.edu/2019/03/23/yes-i-really-really-really-like-fake-data-simulation-and-i-cant-stop-talking-about-it/">fake-data simulation</a> or simulated-data experimentation. I use the term “preregistration” above in order to connect with the many people in the science-reform movement who use that term.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/16/why-i-like-preregistration-and-its-not-about-p-hacking-when-done-right-it-unifies-the-substance-of-science-with-the-scientific-method/feed/</wfw:commentRss>
			<slash:comments>24</slash:comments>
		
		
			</item>
		<item>
		<title>&#8220;The terror among academics on the covid origins issue is like nothing we&#8217;ve ever seen before&#8221;</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/15/the-terror-among-academics-on-the-covid-origins-issue-is-like-nothing-weve-ever-seen-before/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/15/the-terror-among-academics-on-the-covid-origins-issue-is-like-nothing-weve-ever-seen-before/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Wed, 15 Jan 2025 14:19:05 +0000</pubDate>
				<category><![CDATA[Bayesian Statistics]]></category>
		<category><![CDATA[Decision Analysis]]></category>
		<category><![CDATA[Public Health]]></category>
		<category><![CDATA[Sociology]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51564</guid>

					<description><![CDATA[Michael Weissman sends along this article he wrote with a Bayesian evaluation of Covid origins probabilities. He writes: It&#8217;s a peculiar issue to work on. The terror among academics on the covid origins issue is like nothing we&#8217;ve ever seen &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/15/the-terror-among-academics-on-the-covid-origins-issue-is-like-nothing-weve-ever-seen-before/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Michael Weissman sends along <a href="https://michaelweissman.substack.com/p/an-inconvenient-probability-v57">this article</a> he wrote with a Bayesian evaluation of Covid origins probabilities.  He writes:</p>
<blockquote><p>It&#8217;s a peculiar issue to work on. The terror among academics on the covid origins issue is like nothing we&#8217;ve ever seen before.</p></blockquote>
<p>I was surprised he was talking about &#8220;terror&#8221; . . .  People sometimes send me stuff about covid origins and it all seems civil enough. I guess I&#8217;m too far out of the loop to have noticed this!  That said, there have been times that I&#8217;ve been attacked for opposing some aspect of the scientific establishment, so I can believe it.</p>
<p>I asked Weissman to elaborate, and he shared some stories:</p>
<blockquote><p>A couple of multidisciplinary researchers from prestigious institutions were trying to write up a submittable paper. They were leaning heavily zoonotic, at least before we talked. They said they didn&#8217;t publish because they could not get any experts to talk with them. They said they prepared formal legal papers guaranteeing confidentiality but it wasn&#8217;t enough. I guess people thought that their zoo-lean was a ruse.</p>
<p>The extraordinarily distinguished computational biologist Nick Patterson tells me that a prospective collaborator cancelled their collaboration because Patterson had blogged that he thought the evidence pointed to a lab leak. It is not normal for a scientist to drop an opportunity to collaborate with someone like Patterson over a disagreement on an unrelated scientific question. You can imagine the effect of that environment on younger, less established scientists.</p>
<p>Physicist Richard Muller at Berkeley tried asking some bio colleague about an origins-related technical issue. The colleague blew him off. Muller asked if a student or postdoc could help. No way- far too risky, would ruin their career. (see around minute 43 <a href="https://www.youtube.com/watch?v=1jXW7Z3h3nY">here</a>)</p></blockquote>
<p>Come to think about it, I got attacked (or, at least, misrepresented) for some of my <a href="http://stat.columbia.edu/~gelman/research/published/specificity.pdf">covid-related research</a> too; the <a href="https://statmodeling.stat.columbia.edu/2024/12/22/stanford-medical-school-professor-misrepresents-what-i-wrote-but-i-kind-of-understand-where-hes-coming-from/">story is here</a>. Lots of aggressive people out there in the academic research and policy communities.</p>
<p>Also, to put this in the context of the onset of covid in 2020, whatever terror we have been facing by disagreeing with powerful people in academia and government is nothing compared to the terror faced by people who were exposed to this new lethal disease.  Covid is now at the level of a bad flu season, so still pretty bad but much less scary than a few years ago.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/15/the-terror-among-academics-on-the-covid-origins-issue-is-like-nothing-weve-ever-seen-before/feed/</wfw:commentRss>
			<slash:comments>104</slash:comments>
		
		
			</item>
		<item>
		<title>Genre fiction:  Some genres are cumulative and some are not.</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/14/genre-fiction-some-genres-are-cumulative-and-some-are-not/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/14/genre-fiction-some-genres-are-cumulative-and-some-are-not/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Tue, 14 Jan 2025 14:58:10 +0000</pubDate>
				<category><![CDATA[Art]]></category>
		<category><![CDATA[Literature]]></category>
		<category><![CDATA[Sociology]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51047</guid>

					<description><![CDATA[I&#8217;ve been reading some books about the history of twentieth-century mystery novels and science fiction stories, and one thing that struck me was that each of these genres had a sense of continuity. If you used a gimmick in a &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/14/genre-fiction-some-genres-are-cumulative-and-some-are-not/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>I&#8217;ve been reading some books about the history of twentieth-century mystery novels and science fiction stories, and one thing that struck me was that each of these genres had a sense of continuity.  If you used a gimmick in a story, then it was &#8220;yours,&#8221; and future authors were kind of obliged to come up with something new or, if they were going to use your idea, they were supposed to give it a new twist or else refer back to your original story.  And there was a sense of progression, as the mystery puzzles became more elaborate and the science fiction scenarios became more deeply realized.</p>
<p>With an expectation of progression comes a fear of stagnation, and I think that one reason that genre fans talk about a &#8220;golden age&#8221; (that would be the 1930s for mysteries or the 1940s for science fiction) is the idea that you can&#8217;t keep coming up with new tricks.  At some point you need to change the rules, which for mysteries included directions such as interacting psychology and sociology (Ross Macdonald, etc.), focusing on local color and how the world really works (John D. Macdonald, George V. Higgins, etc.), and for science fiction meant a move away from the poles of horror on one side and techno-optimism on the other.  Even as the genres expanded, though, I think there remained a sense of them as cumulative, with new writers building upon what had been done in the past.  You weren&#8217;t supposed to write a mystery novel or science fiction story where you ripped off some previously-published plots without adding something.</p>
<p>What about other genres?</p>
<p>Let&#8217;s start with &#8220;mundane&#8221; or &#8220;memetic&#8221; fiction&#8212;that is, non-genre writing that follows conventions of realism.  There, it&#8217;s considered ok to reuse plots, sometimes very openly as with Jane Smiley&#8217;s remake of King Lear, other times just with standard plot structures of happy families and unhappy families and affairs and business reversals and all sorts of other stories.  In memetic fiction, the plot is not the main focus, and even if the plot is a driver of the story (as with Jonathan Franzen, for example), nobody would really care if it&#8217;s taken from somewhere else.</p>
<p>Other genres commonly mentioned in the same breath as mystery and science fiction are romance, western, porn, and men&#8217;s adventures (war stories, etc.).  I don&#8217;t know much about these genres, so maybe readers can correct me on this, but it&#8217;s my impression that the twentieth-century versions of these genres have not been cumulative in the way of mystery and science fiction.  It was not expected that a romance story or a war story would need a new plot twist or a new idea.  Westerns might be different just because they were so popular for awhile that maybe their high profile pushed authors to come up with new twists, I&#8217;m not sure.  I&#8217;m not saying these genres are static&#8212;they will change over time as readers&#8217; expectations change&#8212;but they&#8217;re not expected to offer novelty or innovation in the way of mystery or science fiction.</p>
<p>Why would these other genres not be cumulative?  Perhaps because they are offering different sorts of pleasures.  Traditionally, mystery and science fiction stories have the form of puzzles; you&#8217;re reading them for the pleasure of trying, and often failing, to figure them out, and then maybe <a href="https://statmodeling.stat.columbia.edu/2024/12/21/alzabo/">rereading</a> for the pleasure of understanding how the mechanisms were put together.  In contrast, memetic fiction and genres such as romance/western/etc. are read more for their emotional impact.  OK, I guess people would also read memetic fiction as a way to learn about the world&#8212;but to learn, you don&#8217;t need a new twist in the story, you just need a clear presentation.</p>
<p>I&#8217;m not saying that mystery and science fiction are read merely for their puzzle aspects.  Mysteries also offer the thrill of suspense and the twin satisfactions of lawbreaking and justice, science fiction has the sense of wonder, and both genres offer some form of social commentary that is gained by looking at society from a distance. I&#8217;m just saying that the puzzle is a big part of Golden Age mystery and science fiction, and this could help explain their cumulative natures.</p>
<p>What other genres could we consider?</p>
<p>There&#8217;s writing for children (recall Orwell&#8217;s classic essay on boys&#8217; weeklies) and young adults, but with rare exceptions these books are read by a new audience every few years, so there&#8217;s no need for continuity.  It&#8217;s no problem if you steal a plot idea from a twenty-year-old book that today&#8217;s kids are no longer reading.</p>
<p>There&#8217;s also modernist fiction, where the innovation is supposed to come in the form, not the content.  You can steal an old plot but you&#8217;re supposed to present it from some new perspective, and that&#8217;s a puzzle in a different way.</p>
<p>So that&#8217;s how I see things as of 1950 or so (with a few forward references).  What&#8217;s been happening since?</p>
<p>So many more books get published each year than before, and nobody can keep track of them.  In the meantime, it seems that fewer people are interested in reading for the puzzle.  Yes, Agatha Christie remains popular, and I&#8217;m guessing that some classic science fiction continues to sell, but I get the impression that, for a long time now, readers of mystery and science fiction aren&#8217;t looking for clever puzzles anymore, with rare exceptions such as Everyone in My Family Has Killed Someone and The Martian.  Mystery and science fiction <em>novels</em> are now more like mystery and science fiction <em>movies</em>, which, again, with rare exceptions, are mostly about delivering thrills, with a side of philosophical reflection.</p>
<p>So, between the disappearance of the past, the diminishing interest in puzzles, and the sheer impossibility of remaining aware of all the earlier books in the genre, I&#8217;m guessing that mystery and science fiction are no longer cumulative endeavors the way they used to be.  Instead of trying to top what came before, their authors are just writing books.</p>
<p>There&#8217;s also the economics of it all:  50 or 100 years ago, you could make an OK living churning out books of any sort, you could make a good living writing successful books, and you had an outside chance of getting rich (by the standards of the day; I&#8217;m talking &#8220;millionaire,&#8221; not &#8220;billionaire&#8221;) by plugging into the zeitgeist and writing bestsellers.  Nowadays, with very rare exceptions, even successful authors don&#8217;t sell many books; many literary authors are reduced to supporting themselves through academic jobs; and pretty much the only way they&#8217;ll make real money from writing is through movie or TV contracts.</p>
<p>To return to the main topic of this post, the transition from a cumulative to a static literature:  this happens in other fields too.  Music, for example,  This happens in different genres at different times, but it seems that in many genres of music, there is a period where different composers and artists are feeding on each others&#8217; work and feel the need to do something new (recall Brian Wilson&#8217;s attitude with respect to the Beatles), and a period where there is no longer the sense of cumulative building of a genre.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/14/genre-fiction-some-genres-are-cumulative-and-some-are-not/feed/</wfw:commentRss>
			<slash:comments>7</slash:comments>
		
		
			</item>
		<item>
		<title>The theory crisis in physics compared to the replication crisis in social science:  Two different opinion-field inversions that differ in some important ways</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/13/the-theory-crisis-in-physics-compared-to-the-replication-crisis-in-social-science-two-different-opinion-field-inversions-that-differ-in-some-important-ways/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/13/the-theory-crisis-in-physics-compared-to-the-replication-crisis-in-social-science-two-different-opinion-field-inversions-that-differ-in-some-important-ways/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Mon, 13 Jan 2025 14:09:15 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Sociology]]></category>
		<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51043</guid>

					<description><![CDATA[Yesterday we discussed an &#8220;opinion-field inversion&#8221; in theoretical physics: in the prestige news media and publicity complex (NPR, Ted, etc), string theory reigns supreme; in elite physics departments, string theory is where it&#8217;s at, but then there&#8217;s a middle range &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/13/the-theory-crisis-in-physics-compared-to-the-replication-crisis-in-social-science-two-different-opinion-field-inversions-that-differ-in-some-important-ways/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><a href="https://statmodeling.stat.columbia.edu/2025/01/12/string-theory-wars-an-opinion-field-inversion/">Yesterday we discussed</a> an &#8220;opinion-field inversion&#8221; in theoretical physics:  in the prestige news media and publicity complex (NPR, Ted, etc), string theory reigns supreme; in elite physics departments, string theory is where it&#8217;s at, but then there&#8217;s a middle range of skeptics ranging from my Columbia colleague Peter &#8220;Not Even Wrong&#8221; Woit to xkcd columnist Randall Munroe who characterize string theory as an overhyped nothingburger.  I guess that Woit would be cool with some elite physicists studying string theory as part of the overall portfolio of theoretical research, but he and others in this middle ground think that whatever profile string theory should have, both in academia and in public perceptions of physics, should be much lower.</p>
<p>I referred to this sort of layered difference in public opinion as an opinion-field inversion, by analogy to the phenomenon of temperature inversion that is a precursor to tornadoes.</p>
<p><strong>The theory crisis in physics</strong></p>
<p>Before going on, let me emphasize that I know nothing of theoretical physics.  I find Woit&#8217;s writing on the topic persuasive, but I&#8217;ve not tried to understand any of the debate in physics terms.  I&#8217;m only discussing this from the perspective of sociology of science.</p>
<p>I&#8217;ll call it the theory crisis in physics, by analogy to the replication crisis in social science and medical research.  The string theory thing isn&#8217;t a replication crisis&#8212;indeed, one of the main criticisms of string theory is that it does not make new testable predictions, so there&#8217;s no possibility of replication or falsification&#8211;but it&#8217;s still a crisis.  I think the term &#8220;theory crisis&#8221; is about right.</p>
<p>Arguably the <a href="http://stat.columbia.edu/~gelman/research/published/jmmss-3062-gelman.pdf">replication crisis in psychology</a> and economics is also a theory crisis, in that the work is based on broad theories such as embodied cognition and evolutionary psychology that have major problems of the sort that they can be used to explain any possible result, but it was the failed replications that were the convincer to many people, hence the term &#8220;replication crisis&#8221; rather than &#8220;theory crisis&#8221; or &#8220;methods crisis.&#8221;</p>
<p><strong>The replication crisis as another opinion-field inversion</strong></p>
<p>In any case, another example of an opinion-field inversion in science, at least until recently, was woo-woo psychology such as social priming and walking speed, ovulation and voting, air rage, power pose, himmicanes, ages ending in 9, signing at the top of the form, etc.  The news media and associated institutions (Ted, Freakonomics, etc.) were all-in on these things; informed scientists such as Uri Simonsohn, Anna Dreber, and various other so-called methodological terrorists were very skeptical; and the power centers at Harvard, PNAS, etc., were a mix of head-in-the-sand true believers (claiming the replication rate &#8220;is statistically indistinguishable from 100%&#8221;) and I&#8217;ve-got-mine-don&#8217;t-rock-the-boat nudgelords, who seemed to be more concerned about keeping their Henry Kissinger party invitations and positions on NPR speed-dial than in cleaning up their house.</p>
<p>With junk science, things have changed&#8211;more and more reporters seem to be tired of having their chains yanked by whatever Psychological Science and PNAS happen to be promoting this week&#8211;and I guess that&#8217;s partly a consequence of the opinion inversion.  Shaking up the power centers might be more of a challenge.</p>
<p><strong>Differences between the theory crisis in physics and the replication crisis in psychology and economics</strong></p>
<p>What will happen with string theory, I don&#8217;t know.  One difference is that in psychology and economics, my impression is that the people who do this sort of headline-bait are not taken very seriously at an intellectual level.  They may have institutional power (hi, Robert Sternberg!) and others in their field may enjoy the reflected glow of their TV appearances, but nobody would consider them to be the brightest lights in the chandelier.  In contrast, it seems that many of the physicists who work in string theory are considered to be brilliant, most notably Ed Witten&#8212;I know nothing of his work, it&#8217;s all beyond me, but he&#8217;s always described in superlatives.</p>
<p>So if string theory really is a hyped dead end, it&#8217;s a much sadder story than junk econ and junk psychology, which seem more like the products of ambitious careerists who, for a couple of decades, stumbled upon a way to hack the system of scientific publication and publicity at the nexus of academia and the news media.</p>
<p>The other twist is that even the opponents of string theory still characterize it as having some mathematical interest&#8211;their criticism is not that string theory is being done, so much with that too much is being claimed for it.  That&#8217;s different than research in himmicanes, air rage, beauty-and-sex-ratio, extra-sensory perception, etc., which is pretty much unmitigated crap, whose only contribution to science has been to reveal the rotten core of science as it is often practiced.</p>
<p>This also gives different flavors to the discussions in these two fields.  With psychology and economics, the frustration is mostly external, with observers being bothered that bad work gets so much publicity, that methodological criticisms and unsuccessful replications get less notice than problematic work, etc.  With physics, the frustration seems mostly internal, with people bothered that the top physics students are going into what they perceive as a dead-end world.  In physics, the concern is with the misuse of human resources in the form of brilliant Ph.D. students.  In psychology and econ, the concern is with the bad work giving the general public a misleading view of science and perhaps leading to bad policies (Excel error, anyone?).  Nobody&#8217;s saying it&#8217;s too bad Brian Wansink and Dan Ariely got into this social priming stuff or otherwise they could&#8217;ve made major discoveries.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/13/the-theory-crisis-in-physics-compared-to-the-replication-crisis-in-social-science-two-different-opinion-field-inversions-that-differ-in-some-important-ways/feed/</wfw:commentRss>
			<slash:comments>47</slash:comments>
		
		
			</item>
		<item>
		<title>String theory wars:  An opinion-field inversion.</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/12/string-theory-wars-an-opinion-field-inversion/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/12/string-theory-wars-an-opinion-field-inversion/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Sun, 12 Jan 2025 14:08:48 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Sociology]]></category>
		<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51042</guid>

					<description><![CDATA[From my Columbia math department colleague Peter Woit: Brian Greene’s The Elegant Universe is being reissued today, in a 25th anniversary edition. It’s the same text as the original, with the addition of a 5 page preface and a 36 &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/12/string-theory-wars-an-opinion-field-inversion/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>From my Columbia math department colleague <a href="https://www.math.columbia.edu/~woit/wordpress/?p=14092">Peter Woit</a>:</p>
<blockquote><p>Brian Greene’s The Elegant Universe is being reissued today, in a 25th anniversary edition. It’s the same text as the original, with the addition of a 5 page preface and a 36 page epilogue. . . .</p>
<p>One thing I [Woit] was looking for in the new material was Greene’s response to the detailed criticisms of string theory that have been made by me and others such as Lee Smolin and Sabine Hossenfelder over the last 25 years. It’s there, and here it is, in full:</p>
<blockquote><p>There is a small but vocal group of string theory detractors who, with a straight face, say things like “A long time ago you string theorists promised to have the fundamental laws of quantum gravity all wrapped up, so why aren’t you done?” or “You string theorists are now going in directions you never expected,” to which I respond, in reverse order “Well, yes, the excitement of searching into the unknown is to discover new directions” and “You must be kidding.”</p></blockquote>
<p>As one of the “small but vocal group” I’ll just point out that this is an absurd and highly offensive straw-man argument. The arguments in quotation marks are not ones being made by string theory detractors, and the fact that he makes up this nonsense and refuses to engage with the real arguments speaks volumes.</p></blockquote>
<p>Woit follows up in the comments section:</p>
<blockquote><p>It’s interesting to see that the comments back up an argument I’ve often heard from physicists when I criticize books like this. They tell me “sure, the material in that book about string theory is nonsense and misleading the public, but it’s getting young people excited about physics and interested in becoming physicists. Once they do become physicists they’ll realize this is nonsense and go on to do some real physics.”</p></blockquote>
<p>Woit&#8217;s a persuasive writer, and he, or someone, <a href="https://xkcd.com/171/">seems to have convinced</a> Randall Munroe.  I have no idea, as my physics days are long gone and I never learned any particle physics in my studies.  When I <a href="http://stat.columbia.edu/~gelman/research/published/2877_001.pdf">worked at the</a> Laboratory for Cosmic Ray Physics we dealt with muons and stuff like that, but nothing so theoretical as string theory or its predecessors.  So I offer no comment on the technical side of all this, except to note that debates remain even in much simpler areas of quantum physics, as I am reminded whenever the discussion comes up of joint distributions in the two-slit experiment (as in section 2 of or article on <a href="http://stat.columbia.edu/~gelman/research/published/physics.pdf">holes in Bayesian statistics</a>): some people will pop up and say I&#8217;m completely wrong, and others will agree with me.  Physics is hard, and even in examples where the theory is well understood, it can be a challenge to come up with a definitive experiment, or even a definitive thought experiment, to resolve fundamental disagreements.</p>
<p>But the sociology-of-science part of the string theory story, that I can comment on.  To start with, Woit and Greene are both in the Columbia math department.  They&#8217;re not far apart in age, and they both studied physics at Harvard.  But, from the quotes above, I get the impression that they can&#8217;t stand each other!  Or, at least, they don&#8217;t respect each other.  That&#8217;s fine&#8212;there&#8217;s no reason the mathematics faculty at Columbia should agree on everything, indeed some healthy disagreement is a good thing.  Still, it&#8217;s interesting.</p>
<p>It&#8217;s also interesting that in some aspects of the court of scientifically-informed public opinion, Woit and the anti-stringers represent the standard or accepted view, whereas Greene and the stringers seem to have locked up the two extremes&#8212;the general public (as represented by PBS, Ted, etc.) and the academic theoretical physics community.</p>
<p>You know how tornados come from &#8220;temperature inversions&#8221;?  I think of this sort of thing as an &#8220;opinion inversion&#8221;:  an unstable pattern in which opinions among more informed people are different from those of the general public.  In this case it&#8217;s a double inversion, with the general public and the institutional power on one side and generally scientifically-informed people in the middle.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/12/string-theory-wars-an-opinion-field-inversion/feed/</wfw:commentRss>
			<slash:comments>82</slash:comments>
		
		
			</item>
		<item>
		<title>Muckraking at the University of Oregon</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/11/muckraking-at-the-university-of-oregon/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/11/muckraking-at-the-university-of-oregon/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Sat, 11 Jan 2025 14:35:35 +0000</pubDate>
				<category><![CDATA[Literature]]></category>
		<category><![CDATA[Political Science]]></category>
		<category><![CDATA[Sociology]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51039</guid>

					<description><![CDATA[Check it out. I was amused by these posts: From 2014: Crap-free UO homepage From 2024: Provost Chris Long is paid $540K plus $130K per year start-up &#038; alcohol budget Columbia should have a crap-free homepage too! And the alcohol &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/11/muckraking-at-the-university-of-oregon/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><a href="https://uomatters.com/">Check it out</a>.  I was amused by these posts:</p>
<p>From 2014:  <a href="https://uomatters.com/uo_links">Crap-free UO homepage</a></p>
<p>From 2024:  <a href="https://uomatters.com/2024/06/provost-chris-long-is-paid-540k-per-year-plus-130k-start-up-alcohol-budget.html">Provost Chris Long is paid $540K plus $130K per year start-up &#038; alcohol budget</a></p>
<p>Columbia should have a crap-free homepage too!  And the alcohol budget reminded me of <a href="https://statmodeling.stat.columbia.edu/2012/12/01/241364-83-13000-228364-83/">this story</a> from Columbia a few years ago.</p>
<p>UO Matters is run by economics professor Bill Harbaugh, who seems to be <a href="https://harbaugh.uoregon.edu">a busy person</a>.  Every city, town, and institution should have this sort of news outlet&#8212;or, ideally, more than one.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/11/muckraking-at-the-university-of-oregon/feed/</wfw:commentRss>
			<slash:comments>24</slash:comments>
		
		
			</item>
		<item>
		<title>&#8220;The king, sir, is much better!&#8221;</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/10/the-king-sir-is-much-better/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/10/the-king-sir-is-much-better/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Fri, 10 Jan 2025 14:23:43 +0000</pubDate>
				<category><![CDATA[Literature]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=50635</guid>

					<description><![CDATA[Algis Budrys wrote this in 1983: Budrys brings this up in the context of how our reading of a book is affected by whatever reviews and publicity materials we&#8217;ve seen&#8211;an interesting point which I became more aware of after going &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/10/the-king-sir-is-much-better/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Algis Budrys <a href="https://books.google.com/books?id=umObBQAAQBAJ&#038;pg=PA28&#038;lpg=PA28&#038;dq=%22the+king,+sir,+is+much+better%22&#038;source=bl&#038;ots=Ps8vPhN-jK&#038;sig=ACfU3U3QkD0u_ndPIp1LUNRerFEH7SJKrA&#038;hl=fr&#038;sa=X&#038;ved=2ahUKEwjl-IaqqKmGAxWUhIkEHeUVAkcQ6AF6BAgIEAM#v=onepage&#038;q&#038;f=false">wrote this in 1983</a>:</p>
<blockquote><p><img fetchpriority="high" decoding="async" src="https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/05/Screenshot-2024-05-25-at-13.24.13-1024x866.png" alt="" width="584" height="494" class="alignnone size-large wp-image-50636" srcset="https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/05/Screenshot-2024-05-25-at-13.24.13-1024x866.png 1024w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/05/Screenshot-2024-05-25-at-13.24.13-300x254.png 300w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/05/Screenshot-2024-05-25-at-13.24.13-768x650.png 768w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/05/Screenshot-2024-05-25-at-13.24.13-355x300.png 355w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/05/Screenshot-2024-05-25-at-13.24.13.png 1518w" sizes="(max-width: 584px) 100vw, 584px" /></p></blockquote>
<p>Budrys brings this up in the context of how our reading of a book is affected by whatever reviews and publicity materials we&#8217;ve seen&#8211;an interesting point which I became more aware of after going to the bookstore in France and picking out books without having a sense of what they&#8217;d be like (see entry under Crédit Illimité in <a href="https://statmodeling.stat.columbia.edu/2023/04/03/48938/">this post</a>.  In any case, the main thing is that I just like Budrys&#8217;s story and how he tells it.  He was a true blogger, which I mean in the best sense of the word.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/10/the-king-sir-is-much-better/feed/</wfw:commentRss>
			<slash:comments>16</slash:comments>
		
		
			</item>
		<item>
		<title>Postdoc, doctoral student, and summer intern positions, Bayesian methods, Aalto</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/10/postdoc-and-doctoral-student-positions-bayesian-methods-aalto/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/10/postdoc-and-doctoral-student-positions-bayesian-methods-aalto/#comments</comments>
		
		<dc:creator><![CDATA[Aki Vehtari]]></dc:creator>
		<pubDate>Fri, 10 Jan 2025 08:30:23 +0000</pubDate>
				<category><![CDATA[Bayesian Statistics]]></category>
		<category><![CDATA[Jobs]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51537</guid>

					<description><![CDATA[Postdoc and doctoral student positions in developing Bayesian methods at Aalto University, Finland! This job post is by Aki (maybe someday I write an actual blog post). The positions are funded by Finnish Center for Artificial Intelligence FCAI and there &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/10/postdoc-and-doctoral-student-positions-bayesian-methods-aalto/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Postdoc and doctoral student positions in developing Bayesian methods at Aalto University, Finland! This job post is by Aki (maybe someday I write an actual blog post).</p>
<p>The positions are funded by <a href="https://fcai.fi/">Finnish Center for Artificial Intelligence FCAI</a> and there are many other topics, but if you specify me as the preferred supervisor then it&#8217;s going to be Bayesian methods, workflow, cross-validation, and diagnostics. See <a href="https://www.youtube.com/watch?v=lKRRyrPxxeU">my video on Bayesian workflow</a>, <a href="https://arxiv.org/abs/2011.01808">Bayesian workflow paper</a>, <a href="https://users.aalto.fi/~ave/publications.html">my publication list</a>, and <a href="https://users.aalto.fi/~ave/videos.html">my talk list</a>, for more about what I&#8217;m working on. </p>
<p>There are also plenty of other topics and supervisors in</p>
<ol>
<li>Reinforcement learning
<li>Probabilistic methods
<li>Simulation-based inference
<li>Privacy-preserving machine learning
<li>Collaborative AI and human modeling
<li>Machine learning for science
</ol>
<p>Join us, and learn why Finland has been ranked the happiest country in the world for seven years in row!</p>
<p>See how to apply at <a href="https://fcai.fi/winter-2025-researcher-positions-in-ai-and-machine-learning">fcai.fi/winter-2025-researcher-positions-in-ai-and-machine-learning</a></p>
<p>I might also hire one summer intern (BSc or MSc level) to work on the same topics. Applications through <a href="https://www.aalto.fi/en/aalto-science-institute-asci/how-to-apply-for-the-asci-international-summer-research-programme">Aalto Science Institute call</a></p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/10/postdoc-and-doctoral-student-positions-bayesian-methods-aalto/feed/</wfw:commentRss>
			<slash:comments>1</slash:comments>
		
		
			</item>
		<item>
		<title>I would add three words to this statement by Uri Simonsohn on preregistration</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/09/i-would-add-three-words-to-this-statement-by-uri-simonsohn-on-preregistration/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/09/i-would-add-three-words-to-this-statement-by-uri-simonsohn-on-preregistration/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Thu, 09 Jan 2025 14:53:01 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Sociology]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51037</guid>

					<description><![CDATA[Uri writes: Pre-registrations should only contain information that helps demarcate confirmatory vs exploratory statistical analyses (i.e., that would help a reader identify harking and p-hacking), and should generally avoid other information. I disagree, partly because I think that confirmatory and &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/09/i-would-add-three-words-to-this-statement-by-uri-simonsohn-on-preregistration/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><a href="http://datacolada.org/119">Uri writes</a>:</p>
<blockquote><p>Pre-registrations should only contain information that helps demarcate confirmatory vs exploratory statistical analyses (i.e., that would help a reader identify harking and p-hacking), and should generally avoid other information.</p></blockquote>
<p>I disagree, partly because I think that confirmatory and exploratory statistical analyses are the same thing (<a href="http://stat.columbia.edu/~gelman/research/published/p755.pdf">see here</a>), partly because I very rarely care about p-values anyway, and mostly because preregistration is super useful to me, not because of harking or p-hacking but because I think that the work we put into preregistration improves our research.  See my discussion <a href="https://statmodeling.stat.columbia.edu/2023/05/29/what-exactly-is-a-preregistration/">here</a> and Jessica&#8217;s <a href="https://statmodeling.stat.columbia.edu/2023/12/04/modest-pre-registration/">here</a>.  For still more, I refer you to my post from 2022, <a href="https://statmodeling.stat.columbia.edu/2022/09/12/whats-the-difference-between-derek-jeter-and-preregistration/">What’s the difference between Derek Jeter and preregistration?</a>.</p>
<p>That said, I can well believe that, <em>for Uri</em>, the best preregistration only contain information that helps demarcate etc etc.</p>
<p>So my suggested alteration to Uri&#8217;s above-quoted statement is just to add &#8220;for Uri Simonsohn&#8221; before the word &#8220;should.&#8221;  Problem solved!</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/09/i-would-add-three-words-to-this-statement-by-uri-simonsohn-on-preregistration/feed/</wfw:commentRss>
			<slash:comments>6</slash:comments>
		
		
			</item>
		<item>
		<title>Junk science becomes more professionalized.  Meanwhile, conspiracy theories are being more associated with the center-right and right, politically.  How does all this fit together?  I&#8217;m not sure.</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/08/junk-science-becomes-more-professionalized-meanwhile-conspiracy-theories-are-being-more-associated-with-the-political-right-how-does-all-this-fit-together-im-not-sure/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/08/junk-science-becomes-more-professionalized-meanwhile-conspiracy-theories-are-being-more-associated-with-the-political-right-how-does-all-this-fit-together-im-not-sure/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Wed, 08 Jan 2025 14:38:48 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Political Science]]></category>
		<category><![CDATA[Sociology]]></category>
		<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51020</guid>

					<description><![CDATA[A few years ago I wrote a post, Junk Science Then and Now discussing the movement of junk science from the periphery of elite culture to the core: The junk science (by which I mean work that has some of &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/08/junk-science-becomes-more-professionalized-meanwhile-conspiracy-theories-are-being-more-associated-with-the-political-right-how-does-all-this-fit-together-im-not-sure/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>A few years ago I wrote a post, <a href="https://statmodeling.stat.columbia.edu/2020/03/06/junk-science-then-and-now/">Junk Science Then and Now</a> discussing the movement of junk science from the periphery of elite culture to the core:</p>
<blockquote><p>The junk science (by which I mean work that has some of the forms of scientific research but is missing key elements such as valid and reliable measurements, transparency, and openness to criticism) of the mid-twentieth century came from cranks and outsiders, often self-educated people with no academic positions, and even those who were in academia were peripheral figures, for example, the ESP researcher J. B. Rhine at Duke University, who according to Wikipedia was trained as a botanist and was not a central figure in the psychology profession. Immanuel “Worlds in Collision” Velikovsky had lots of scientist friends, but he was an outsider to the scientific community. And those guys from the 1970s who wrote books about ancient astronauts and the Bermuda triangle, I don’t think they even claimed to have any scientific backing. Yes, there were some missteps within academic science from N-rays to cold fusion, but these were minor storms that blew up and went away.</p>
<p>Nowadays, though, the pseudoscientists are well ensconced in the academy, they play power games in the field of psychology, and they get to publish in the Proceedings of the National Academy of Sciences (air rage, himmicanes, ages ending in 9, etc etc) whenever they want. The call is coming from inside the house, as it were. Many of them are still considered by the news media to be the legitimate representatives of the scientific community.  Even absolutely ridiculous ideas like the <a href="https://statmodeling.stat.columbia.edu/2021/09/21/more-on-that-claim-that-scientific-citations-are-worth-100000-each/">$100,000 citations</a>.  There&#8217;s also the related phenomenon of . . . not &#8220;junk science&#8221; exactly, but bad science:  scientific errors that then persist because the scientific community refuses to come to terms with corrections.  An example is the <a href="https://statmodeling.stat.columbia.edu/2024/09/07/christakis-fowler-update-update/">contagion-of-obesity</a> story.</p></blockquote>
<p>When discussing this, I wrote that the above-described shift represents a sort of gentrification of scientific error, mirroring the professionalism that has come into so many other aspects of our intellectual life.  Instead of some wacky guy somewhere claiming to have developed a perpetual motion machine or whatever, you&#8217;ve got a <a href="https://statmodeling.stat.columbia.edu/2023/07/10/here-are-the-data-from-that-cold-showers-study-so-you-haters-can-now-do-your-own-analyses/">Stanford professor</a> promoting junk science on cold showers.</p>
<p>I thought about all this recently when reading <a href="https://www.slowboring.com/p/the-crank-realignment-is-bad-for">a post</a> by political journalist Matthew Yglesias on what he calls &#8220;the crank realignment&#8221;:</p>
<blockquote><p>Robert F. Kennedy Jr.’s transition from semi-prominent Democrat to third party spoiler to Donald Trump endorser is emblematic of a broader, decade-long “crank realignment” in American politics.</p>
<p>Trump himself, of course, used to be a Democrat. He switched parties in a blaze of birther conspiracy theories, and only then came to embrace conservative views on topics like gun control and abortion. And RFK Jr. was into election fraud conspiracy theories long before January 6, but his version was about George W. Bush stealing the 2004 election in Ohio. That wasn’t a mainstream Democratic Party view (there’s a reason there was no Kerry-led insurrection), but it was mainstream enough to be published in Rolling Stone and for Kennedy to continue to be a player in progressive politics.</p>
<p>Twenty years later, that’s no longer the case. Democrats are much more buttoned-up, and the GOP is much more accepting of cranks and know-nothings like Kennedy.</p>
<p>The partisan shifts of both Trump and RFK Jr. are part of a long term cycle in which educated professionals have gravitated toward the Democratic Party coalition and a generic suspicion of institutions and the people who run them has come to be associated with conservative politics.</p></blockquote>
<p>I think Yglesias is on to something here.  I agree with him that from a logical point of view, there&#8217;s no reason why conspiracy theories should be concentrated on the right half of the political spectrum.  Indeed, from a logical perspective you might expect conspiracy theories to be more popular on the left, as this would be consistent with a general leftist anti-powerful-people, anti-big-business take.</p>
<p>One thing I&#8217;ve noticed in the past is that commentators have been tied to the idea of anti-science leftists even when the data don&#8217;t bear that out.  <a href="https://statmodeling.stat.columbia.edu/2022/05/26/public-opinion-isnt-always-as-polarized-as-you-think-gmo-food-edition/">Here&#8217;s an example</a> from a couple years ago, where political scientist Chris Blattman made the offhanded remark that opposition to genetically modified organisms (GMOs) was &#8220;mostly left,&#8221; even though actually opposition was about the same on the left and right.  It&#8217;s a convenient story to pair anti-vaccine attitudes on the right to anti-GMO attitudes on the left and ask why can&#8217;t we all get along, but that&#8217;s not what public opinion happens to look like.  I agree with Yglesias that this is kinda too bad, as it makes it harder to have a bipartisan push against anti-science.</p>
<p>It&#8217;s still hard for me to put all this together in my head.  One challenge is that I have the impression that most of the prominent purveyors of junk science and bad science in academia are on the left, or the center-left.  OK, not Dr. Oz, and maybe not that cold-shower dude at Stanford.  And not those <a href="https://statmodeling.stat.columbia.edu/2021/01/29/team-stanford/">covid-minimizers</a>.  Or the climate-change denialists being promoted by Freakonomics.  But the mainstream NPR/Ted/PNAS world . . . they&#8217;re mostly on the left, right?  We do hear about right-wing people in science, but they get some attention because they are exceptions.</p>
<p>So the professionalization of bullshit&#8212;as exemplified by Gladwell&#8217;s prominence at the New Yorker, the <a href="https://statmodeling.stat.columbia.edu/2024/06/04/again-on-the-role-of-elite-media-in-spreading-ufos-as-space-aliens-and-other-bad-ideas/">UFO&#8217;s-as-space-aliens</a> theories promoted by elite journalists, the Association for Psychological Science <a href="https://statmodeling.stat.columbia.edu/2021/12/20/not-replicable-but-citable/">promotion of superstition</a>, and various wacky stuff coming out of Harvard, Stanford, etc.&#8212;runs counter to the movement of conspiracy theorizing from the political fringes to the core of the political right.</p>
<p>I don&#8217;t know where this will all lead.  It seems kind of unstable.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/08/junk-science-becomes-more-professionalized-meanwhile-conspiracy-theories-are-being-more-associated-with-the-political-right-how-does-all-this-fit-together-im-not-sure/feed/</wfw:commentRss>
			<slash:comments>36</slash:comments>
		
		
			</item>
		<item>
		<title>Treasure trove of forensic details in arXiv&#8217;s LaTeX source code</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/07/treasure-trove-of-forensic-details-in-arxivs-latex-source-code/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/07/treasure-trove-of-forensic-details-in-arxivs-latex-source-code/#comments</comments>
		
		<dc:creator><![CDATA[Bob Carpenter]]></dc:creator>
		<pubDate>Tue, 07 Jan 2025 20:00:25 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Stan]]></category>
		<category><![CDATA[Statistical Computing]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51520</guid>

					<description><![CDATA[There&#8217;s gold in them thar hills* When you submit a paper to arXiv, you send them a bundle including the LaTeX source, figures, etc. These are all available for download through the arXiv site. This morning, I was downloading the &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/07/treasure-trove-of-forensic-details-in-arxivs-latex-source-code/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><b>There&#8217;s gold in them thar hills<sup>*</sup></b></p>
<p>When you submit a paper to <i>arXiv</i>, you send them a bundle including the LaTeX source, figures, etc.  These are all available for download through the <i>arXiv</i> site.  This morning, I was downloading the source<sup>**</sup> for the original Hoffman and Gelman no-U-turn sampler paper.  If you want to follow along, ere&#8217;s the <a href="https://arxiv.org/abs/1111.4246">arXiv link</a>, but you have to click through to the &#8220;TeX Source&#8221; link under the &#8220;Access Paper:&#8221; header on the top right side under the banner.  What I found was a treasure trove of comments that never made it to the paper, some of which I will share below.</p>
<p><b>Examples</b></p>
<p>Returning to Hoffman and Gelman&#8217;s <i>arXiv</i> source LaTeX, what struck me was the following comment right after the algorithm itself.</p>
<pre>
%% Algorithm ?? is more efficient than algorithm
%% ??, but the policy of sampling uniformly from
%% $\cC$ leaves something to be desired. We would prefer to select an
%% element of $\cC$ that is farther away from the initial position
%% $\theta^t$, rather than face the possibility of performing many costly
%% gradient evaluations just to wind up choosing an element of $\cC$ that
%% is close to where we started. Algorithm ?? addresses
%% this issue by giving preference to points subtrees that do not include
%% the starting point $\{\theta^t, r^t\}$. [To do: explain why this is
%%   valid. Probably proof by induction is the easiest way to go.]
</pre>
<p>Neither the <i>arXiv</i> preprint nor the final <i>JMLR</i> paper have a clearly delineated inductive proof.  In both versions, we get &#8220;this is equivalent to a Metropolis_Hastings kernel with proposal &#8230;, and it is straightforward to show that it obeys detailed balance&#8221; (second-to-last sentence on page 1604 of <i>JMLR</i> paper).</p>
<p>Presumably on the principle of minimizing surface area for reviewers to gripe, the following useful comment from the abstract didn&#8217;t make the final cut.</p>
<pre>
%% This issue is compounded when the
%% target distribution depends on a set of parameters that cannot be
%% updated by HMC (such as discrete parameters) and are updated
%% independently of the parameters updated by HMC. 
%% In this case, optimal settings of $L$ may change from iteration to
%% iteration.
</pre>
<p>Here&#8217;s another useful comment that wound up on the cutting-room floor.  I&#8217;m not saying these should all be in the paper&#8212;usually there are so many things you can add and qualify that it requires some judgement.  But for the dedicated and interested reader, the paper would have been more useful with the elided comments.</p>
<pre>
%% Even if we assume that there exists some transformation of the
%% parameter space under which all parameters are i.i.d. and that this
%% transformation can be applied cheaply (i.e. in $O(D)$ time, for
%% example using a low-rank transformation matrix to avoid the $O(D^2)$
%% cost of dense matrix multiplication and the $O(D^3)$ cost of dense
%% matrix inversion), the cost of obtaining an effectively independent
%% sample using RWM is still $O(D^2)$ \citep{Creutz:1988}. Gibbs also
%% requires $O(D^2)$ operations per effectively independent sample in
%% this setting, since it must update $D$ parameters and it must perform
%% a transformation costing $O(D)$ operations after each update.
</pre>
<p>There are also useful explanations of figures that never made the final cut, like this one, which expands the diagram from the one of &#8220;naive NUTS&#8221; in the paper figures to what the paper calls &#8220;efficient NUTS.&#8221;</p>
<pre>
%% %% %% Figure ?? illustrates how an iteration of NUTS might
%% %% %% proceed once the slice and initial momentum variables have been
%% %% %% resampled. Initially (a), we have only one node. We double the size of
%% %% %% the tree to two nodes by taking a single step forward (b), and since
%% %% %% the new point is valid tentatively set $w^{t+1}$ to that new point
%% %% %% (with probability $1/1=1$). We then redouble the size of the tree to
%% %% %% four nodes, taking two steps forward (c). Only one of the two new
%% %% %% nodes is valid, so the probability of choosing a node from the new
%% %% %% half-tree is $1/2$ (the ratio of the number of valid new nodes to
%% %% %% valid old nodes). In this example, we randomly choose to stick with
%% %% %% the old value of $w^{t+1}$. Next, we again double the size of the tree
%% %% %% by taking four steps backward from $w^-$ (d). We discover that the new
%% %% %% half-tree satisfies the stopping criterion, and so we cannot select
%% %% %% any points from it. Finally, we double the tree one more time, this
%% %% %% time going forward (e). This half-tree contains some valid points and
%% %% %% does not satisfy the stopping criterion, but a subtree of it does
%% %% %% satisfy the stopping criterion, so we invalidate the points in that
%% %% %% subtree. The number of valid points in this half-tree (3) is the same
%% %% %% as the number of valid points in the old half-tree (3), so we choose a
%% %% %% point uniformly at random from the new half-tree for $w^{t+1}$. At
%% %% %% this point, the end points $w^-$ and $w^+$ satisfy the stopping
%% %% %% criterion, and we return $w^{t+1}$ as the new position-momentum pair.
</pre>
<p>The following would have been nice.</p>
<pre>
%% Also, we should probably have a scatterplot showing target versus
%% realized criteria (mean acceptance probability, mean energy change)
%% that shows that the stochastic approximation scheme pretty much works,
%% and maybe a plot showing convergence speed.
</pre>
<p>There&#8217;s more where these came from&#8212;I was just cherry picking from the algorithm, abstract, intro, and conclusion.</p>
<p><b>Who knew?</b></p>
<p>I&#8217;ve never heard anyone mention diving into the source of papers, so I wonder just what&#8217;s out there to be mined.  I also wonder how many authors realize that comments in their arXiv LaTeX are forever.</p>
<hr />
<p><small><sup>*</sup> An American idiom meaning there&#8217;s value to be found from exploring in a particular place; see <a href="https://en.wiktionary.org/wiki/gold_in_them_thar_hills">the wikitionary</a> for a definition and etymology.</small></p>
<p><small><sup>**</sup> I downloaded the LaTeX source of Hoffman and Gelman&#8217;s paper in order to produce a <a href="https://chatgpt.com/share/6772fadf-ac88-8011-a6d6-7e83539707d3">ChatGPT(o1[plus]) translation</a> of the efficient NUTS algorithm to Python.  I need to code a similar algorithm for a new sampler we&#8217;re exploring and wanted to make sure I had understood the structure of the NUTS algorithm, because it&#8217;s a very subtle recursion.  GPT continues to impress!<br />
</small></p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/07/treasure-trove-of-forensic-details-in-arxivs-latex-source-code/feed/</wfw:commentRss>
			<slash:comments>3</slash:comments>
		
		
			</item>
		<item>
		<title>Truth is more realistic than fiction, and what this tells us about odious thought experiments</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/07/truth-is-more-realistic-than-fiction-and-what-this-tells-us-about-odious-thought-experiments/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/07/truth-is-more-realistic-than-fiction-and-what-this-tells-us-about-odious-thought-experiments/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Tue, 07 Jan 2025 14:42:37 +0000</pubDate>
				<category><![CDATA[Political Science]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51021</guid>

					<description><![CDATA[In 2010, economist Robin Hanson gained some notoriety by writing about &#8220;gentle silent rape&#8221;: Imagine a woman was drugged into unconsciousness and then gently raped, so that she suffered no noticeable physical harm nor any memory of the event, and &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/07/truth-is-more-realistic-than-fiction-and-what-this-tells-us-about-odious-thought-experiments/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>In 2010, economist Robin Hanson gained some notoriety by <a href="https://www.overcomingbias.com/p/gentlesilentrapehtml">writing about</a> &#8220;gentle silent rape&#8221;:</p>
<blockquote><p>Imagine a woman was drugged into unconsciousness and then gently raped, so that she suffered no noticeable physical harm nor any memory of the event, and the rapist tried to keep the event secret. Now drugging someone against their will is a crime, but the added rape would add greatly to the crime in the eyes of today’s law, and the added punishment for this addition would be far more than for cuckoldry. . . . A colleague of mine suggests this is gender bias, pure and simple; women seem feminist, and men chivalrous, by railing against rape, but no one looks good complaining about cuckoldry. What other explanations you got?</p></blockquote>
<p>Hanson&#8217;s wikipedia entry contains this quote from Nate Silver from 2012:</p>
<blockquote><p>He is clearly not a man afraid to challenge the conventional wisdom. Instead, Hanson writes a blog called Overcoming Bias, in which he presses readers to consider which cultural taboos, ideological beliefs, or misaligned incentives might constrain them from making optimal decisions.</p></blockquote>
<p>Taking these phrases one at time:</p>
<p>&#8211; &#8220;Cultural taboos&#8221; = attitudes that the many people have but which you don&#8217;t share.<br />
&#8211; &#8220;Ideological beliefs&#8221; = ideologies that many people hold that you don&#8217;t share.<br />
&#8211; &#8220;Misaligned incentives&#8221; = incentives for people to do things that make you unhappy.<br />
&#8211; &#8220;Optimal decision&#8221; = decisions that you approve of.</p>
<p>In any case, the &#8220;gentle silent rape&#8221; thing sounded like a <a href="https://slate.com/business/2018/04/economist-robin-hanson-might-be-americas-creepiest-professor.html">bizarre</a> thought experiment.  But then <a href="https://www.nytimes.com/2024/09/05/world/europe/france-rape-trial-pelicot-testimony.html">a news item recently appeared</a> in which it really happened:  a man had drugged and raped his wife and kept it secret for decades.  The result was neither gentle nor silent, which I guess might lead Hanson to say that this real-world case wasn&#8217;t an example of what he was talking about, but I would take this argument in the opposite direction and say that the real-world horror story demonstrates a problem with the thought experiment, which is that gentle silent rape isn&#8217;t really a thing&#8212;the phrase is a way of minimizing a real crime by giving it impossible modifying qualifiers.</p>
<p>The point of this post is not to have some sort of gotcha on Nate.  Rather, it&#8217;s just a horribly vivid demonstration of a general issue with thought experiments in social science, which is that to work they should be internally coherent and also consistent with reality.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/07/truth-is-more-realistic-than-fiction-and-what-this-tells-us-about-odious-thought-experiments/feed/</wfw:commentRss>
			<slash:comments>49</slash:comments>
		
		
			</item>
		<item>
		<title>Progress in 2024 (Aki)</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/07/progress-in-2024-aki/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/07/progress-in-2024-aki/#respond</comments>
		
		<dc:creator><![CDATA[Aki Vehtari]]></dc:creator>
		<pubDate>Tue, 07 Jan 2025 08:30:53 +0000</pubDate>
				<category><![CDATA[Bayesian Statistics]]></category>
		<category><![CDATA[Miscellaneous Statistics]]></category>
		<category><![CDATA[Statistical Computing]]></category>
		<category><![CDATA[Teaching]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51530</guid>

					<description><![CDATA[Here&#8217;s my 2024 progress report. There are 5 publications common with Andrew in 2024. Active Statistics book is the biggest in size, but personally getting the Pareto smoothed importance sampling paper published after 9 years from the first submission was &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/07/progress-in-2024-aki/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Here&#8217;s my 2024 progress report. There are <a href="https://statmodeling.stat.columbia.edu/2025/01/01/published-or-accepted-for-publication-in-2024/">5 publications common with Andrew in 2024</a>.</p>
<p>Active Statistics book is the biggest in size, but personally getting the Pareto smoothed importance sampling paper published after 9 years from the first submission was a big event, too. I think I only blogged 2023 progress report and job ads (I sometimes have blog post ideas, but as I&#8217;m a slow writer, it&#8217;s difficult to find time to turn them to actual posts). I&#8217;m very happy with the progress in 2024, but also excited on what we are going to get done in 2025!</p>
<h2>Book</h2>
<ul>
<li> Andrew Gelman and Aki Vehtari (2024). <a href="https://avehtari.github.io/ActiveStatistics/">Active Statistics</a>.
</ul>
<h2>Papers published or accepted for publication in 2024</h2>
<ul>
<li>
<p>Yann McLatchie, Sölvi Rögnvaldsson, Frank Weber, and Aki Vehtari (2025). Advances in projection predictive inference. <em>Statistical Science</em>, accepted for publication. <a href="https://arxiv.org/abs/2306.15581">arXiv preprint arXiv:2306.15581</a>. Software: <a href="https://mc-stan.org/projpred/">projpred</a>, <a href="https://kulprit.readthedocs.io/">kulprit</a>.</p>
</li>
<li>
<p>Christopher Tosh, Philip Greengard, Ben Goodrich, Andrew Gelman, Aki Vehtari, and Daniel Hsu (2025). The piranha problem: Large effects swimming in a small pond. <em>Notices of the American Mathematical Society</em>, 72(1):15-25. <a href="https://arxiv.org/abs/2105.13445">arXiv preprint arXiv:2105.13445</a>.</p>
</li>
<li>
<p>Kunal Ghosh, Milica Todorović, Aki Vehtari, and Patrick Rinke (2025). Active learning of molecular data for task-specific objectives. <em>The Journal of Chemical Physics</em>, <a href="https://doi.org/10.1063/5.0229834">doi:10.1063/5.0229834</a>.</p>
</li>
<li>
<p>Charles C. Margossian, Matthew D. Hoffman, Pavel Sountsov, Lionel Riou-Durand, Aki Vehtari, and Andrew Gelman (2024). Nested Rhat: Assessing the convergence of Markov chain Monte Carlo when running many short chains. <em>Bayesian Analysis</em>, <a href="https://doi.org/10.1214/24-BA1453">doi:10.1214/24-BA1453</a>.Software: <a href="https://mc-stan.org/posterior/">posterior</a>.</p>
</li>
<li>
<p>Yann McLatchie and Aki Vehtari (2024). Efficient estimation and correction of selection-induced bias with order statistics. <em>Statistics and Computing</em>, 34(132). <a href="https://doi.org/10.1007/s11222-024-10442-4">doi:10.1007/s11222-024-10442-4</a>.</p>
</li>
<li>
<p>Frank Weber, Änne Glass, and Aki Vehtari (2024). Projection predictive variable selection for discrete response families with finite support. <em>Computational Statistics</em>, <a href="https://doi.org/10.1007/s00180-024-01506-0">doi:10.1007/s00180-024-01506-0</a>. Software <a href="https://mc-stan.org/projpred/">projpred</a>.</p>
</li>
<li>
<p>Aki Vehtari, Daniel Simpson, Andrew Gelman, Yuling Yao, and Jonah Gabry (2024). Pareto smoothed importance sampling. <em>Journal of Machine Learning Research</em>, 25(72):1-58. <a href="https://jmlr.org/papers/v25/19-556.html">Online</a>. Software: <a href="https://mc-stan.org/loo/">loo</a>, <a href="https://mc-stan.org/posterior/">posterior</a>, <a href="https://python.arviz.org/">ArviZ</a></p>
</li>
<li>
<p>Manushi Welandawe, Michael Riis Andersen, Aki Vehtari, and Jonathan H. Huggins (2024). A framework for improving the reliability of black-box variational inference. <em>Journal of Machine Learning Research</em>, 25(219):1-71. <a href="https://jmlr.org/papers/v25/22-0327.html">Online</a>.</p>
</li>
<li>
<p>Noa Kallioinen, Topi Paananen, Paul-Christian Bürkner, and Aki Vehtari (2024). Detecting and diagnosing prior and likelihood sensitivity with power-scaling. <em>Statistics and Computing</em>, 34(57). <a href="https://doi.org/10.1007/s11222-023-10366-5">Online</a>.<br /> <a href="https://github.com/n-kall/powerscaling-sensitivity">Supplementary code</a>.<br /> Software: <a href="https://github.com/n-kall/priorsense">priorsense</a></p>
</li>
<li>
<p>Erik Štrumbelj, Alexandre Bouchard-Côté, Jukka Corander, Andrew Gelman, Håvard Rue, Lawrence Murray, Henri Pesonen, Martyn Plummer, and Aki Vehtari (2024). Past, present, and future of software for Bayesian inference. <em>Statistical Science</em>, 39(1):46-61. <a href="https://doi.org/10.1214/23-STS907">Online</a>.</p>
</li>
<li>
<p>Alex Cooper, Dan Simpson, Lauren Kennedy, Catherine Forbes, and Aki Vehtari (2024). Cross-validatory model selection for Bayesian autoregressions with exogenous regressors. <em>Bayesian Analysis</em>, <a href="https://doi.org/10.1214/23-BA1409">doi:10.1214/23-BA1409</a>.</p>
</li>
<li>
<p>Marta Kołczyńska, Paul-Christian Bürkner, Lauren Kennedy, and Aki Vehtari (2024). Trust in state institutions in Europe, 1989–2019. <em>Survey Research Methods</em>, 18(1). <a href="https://doi.org/10.18148/srm/2024.v18i1.8119">doi:10.18148/srm/2024.v18i1.8119</a>.</p>
</li>
<li>
<p>Alex Cooper, Aki Vehtari, Catherine Forbes, Lauren Kennedy, and Dan Simpson (2024). Bayesian cross-validation by parallel Markov chain Monte Carlo. <em>Statistics and Computing</em>, 34:119. <a href="https://doi.org/10.1007/s11222-024-10404-w">doi:10.1007/s11222-024-10404-w</a>.</p>
</li>
<li>
<p>Ryoko Noda, Michael Francis Mechenich, Juha Saarinen, Aki Vehtari, Indrė Žliobaitė (2024). Predicting habitat suitability for Asian elephants in non-analog ecosystems with Bayesian models. <em>Ecological Informatics</em>, 82:102658. <a href="https://doi.org/10.1016/j.ecoinf.2024.102658">doi:10.1016/j.ecoinf.2024.102658</a>.</p>
</li>
<li>
<p>Petrus Mikkola, Osvaldo A. Martin, Suyog Chandramouli, Marcelo Hartmann, Oriol Abril Pla, Owen Thomas, Henri Pesonen, Jukka Corander, Aki Vehtari, Samuel Kaski, Paul-Christian Bürkner, Arto Klami (2024). Prior knowledge elicitation: The past, present, and future. <em>Bayesian Analysis</em>, 19(49):1129-1161. <a href="https://doi.org/10.1214/23-BA1381">doi:10.1214/23-BA1381</a>.</p>
</li>
</ul>
<h2>arXived in 2024</h2>
<ul>
<li>
<p>Marvin Schmitt, Chengkun Li, Aki Vehtari, Luigi Acerbi, Paul-Christian Bürkner, and Stefan T. Radev (2024). Amortized Bayesian Workflow (Extended Abstract). <a href="https://arxiv.org/abs/2409.04332">arXiv preprint arXiv:2409.04332</a>.</p>
</li>
<li>
<p>Måns Magnusson, Jakob Torgander, Paul-Christian Bürkner, Lu Zhang, Bob Carpenter, and Aki Vehtari (2024). posteriordb: Testing, benchmarking and developing Bayesian inference algorithms. <a href="https://arxiv.org/abs/2407.04967">arXiv preprint arXiv:2407.04967</a>. Database and software: <a href="https://github.com/stan-dev/posteriordb">posteriordb</a></p>
</li>
<li>
<p>David Kohns, Noa Kallionen, Yann McLatchie, and Aki Vehtari (2024). The ARR2 prior: flexible predictive prior definition for Bayesian auto-regressions. <a href="https://arxiv.org/abs/2405.19920">arXiv preprint arXiv:2405.19920</a>.</p>
</li>
<li>
<p>Anna Elisabeth Riha, Nikolas Siccha, Antti Oulasvirta, and Aki Vehtari (2024). Supporting Bayesian modelling workflows with iterative filtering for multiverse analysis. <a href="https://arxiv.org/abs/2404.01688">arXiv preprint arXiv:2404.01688</a>.</p>
</li>
<li>
<p>Guangzhao Cheng, Aki Vehtari, and Lu Cheng (2024). Raw signal segmentation for estimating RNA modifications and structures from Nanopore direct RNA sequencing data. <a href="https://doi.org/10.1101/2024.01.11.575207">bioRxiv preprint</a>.</p>
</li>
</ul>
<h2>Software</h2>
<ul>
<li>
<p><a href="https://mc-stan.org/about/team/">Stan development team</a>. <a href="https://mc-stan.org/">Stan</a>.<br />Releases <a href="https://blog.mc-stan.org/2024/01/16/release-of-cmdstan-2-34/">v2.34</a>, <a href="https://blog.mc-stan.org/2024/06/03/release-of-cmdstan-2-35/">v2.35</a>, <a href="https://blog.mc-stan.org/2024/12/10/release-of-cmdstan-2-36/">v2.36</a>.</p>
</li>
<li>
<p>Aki Vehtari, Jonah Gabry, Måns Magnusson, Yuling Yao, Paul-Christian Bürkner, Topi Paananen, and Andrew Gelman (2024). loo: Efficient Leave-One-Out Cross-Validation and WAIC for Bayesian Models. <a href="https://mc-stan.org/loo/">mc-stan.org/loo/</a>. <br /><a href="https://github.com/stan-dev/loo/releases">Releases v2.7.0, v2.8.0</a></p>
</li>
<li>
<p>Paul-Christian Bürkner, Jonah Gabry, Matthew Kay, and Aki Vehtari (2024). posterior: Tools for Working with Posterior Distributions. <a href="https://mc-stan.org/posterior/">mc-stan.org/posterior/</a>.<br />
  <a href="https://github.com/stan-dev/posterior/releases/tag/v1.6.0">Release 1.6.0</a></p>
</li>
<li>
<p>Noa Kallioinen, Paul-Christian Bürkner, Topi Paananen, Frank Weber, and Aki Vehtari (2024). priorsense: Detecting and diagnosing prior and likelihood sensitivity with power-scaling. <a href="https://github.com/n-kall/priorsense">github.com/n-kall/priorsense</a>.<br />
  <a href="https://cran.r-project.org/web/packages/priorsense/news/news.html">Release 1.0.*</a> (first CRAN release!)</p>
</li>
</ul>
<h2>Case studies</h2>
<ul>
<li>
<p>Aki Vehtari (2024). <a href="https://users.aalto.fi/~ave/casestudies/Nabiximols/nabiximols.html">Nabiximols</a>. Model checking and comparison, comparison of continuous and discrete models, LOO-PIT checking, calibration plots, prior sensitivity analysis, model refinement, treatment effect, effect of model mis-specification.</p>
</li>
<li>
<p>Aki Vehtari (2024). <a href="https://users.aalto.fi/~ave/casestudies/Birthdays/birthdays.html">Birthdays</a>. Workflow example for iterative building of a time series model. In 2024, added demonstration of <a href="https://jmlr.org/papers/v23/21-0889.html">Pathfinder</a> for quick initial results and MCMC initialization.</p>
</li>
</ul>
<h2>FAQ</h2>
<ul>
<li>
<p>Aki Vehtari (2024). <a href="https://users.aalto.fi/~ave/CV-FAQ.html">Cross-validation FAQ</a>. Updates.</p>
</li>
</ul>
<h2>Video</h2>
<ul>
<li>
<p>Aki Vehtari (2024). <a href="https://www.youtube.com/watch?v=12OMXQFbW6I">Pareto-k diagnostic and sample size needed for CLT to hold</a> (StanCon 2024)</p>
</li>
</ul>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/07/progress-in-2024-aki/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>This looks like an excellent new business line for Wolfram Research!</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/06/literary-agency-scams/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/06/literary-agency-scams/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Mon, 06 Jan 2025 14:52:25 +0000</pubDate>
				<category><![CDATA[Economics]]></category>
		<category><![CDATA[Literature]]></category>
		<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51009</guid>

					<description><![CDATA[Can you catch the trick in the above letter? I was staring and staring and couldn&#8217;t figure out the scam. Yes, I get it that &#8220;Mushens and Churchill&#8221; is a fake literary agency (according to this post from Victoria Strauss, &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/06/literary-agency-scams/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><img decoding="async" src="https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/image.png" alt="" width="852" height="813" class="alignnone size-full wp-image-51010" srcset="https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/image.png 852w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/image-300x286.png 300w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/image-768x733.png 768w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/image-314x300.png 314w" sizes="(max-width: 852px) 100vw, 852px" /></p>
<p>Can you catch the trick in the above letter?  I was staring and staring and couldn&#8217;t figure out the scam.  Yes, I get it that &#8220;Mushens and Churchill&#8221; is a fake literary agency (according to <a href="https://writerbeware.blog/2024/08/16/the-latest-fake-literary-agencies/">this post from Victoria Strauss</a>, which is where I found this story, this particular scammer is taking the name of the legitimate literary agent Juliet Mushens, which is a really horrible thing to do), and they&#8217;re preying on the hopes of authors.  I get that &#8220;we cannot promise the moon and the stars&#8221; is classic soft-sell.  What I couldn&#8217;t figure out is what&#8217;s the motivation for the scammer.  They get someone to send them their unpublished manuscript?  Later they ask for money to publish the book?  But that doesn&#8217;t make sense&#8212;the author is already self-publishing.  (And, yes, there&#8217;s no shame in paying money to publish your own work&#8212;do you think this blog hosting comes for free?)</p>
<p><a href="https://writerbeware.blog/2024/08/16/the-latest-fake-literary-agencies/">Strauss explains</a> how the scam works:</p>
<blockquote><p>Although I haven’t yet heard from anyone who has actually signed up with MCLit, and therefore don’t know what they’re charging, the fifth paragraph of the solicitation above gives away what they’re selling: an “International Literary Registration Seal and Bookstore Access Code”. Both of these are completely bogus items that scammers have invented to enable them to drain writers’ bank accounts.</p></blockquote>
<p>Ha!  I didn&#8217;t catch that at all.</p>
<p>Here&#8217;s another:</p>
<p><img decoding="async" src="https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/image-1.png" alt="" width="866" height="956" class="alignnone size-full wp-image-51011" srcset="https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/image-1.png 866w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/image-1-272x300.png 272w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/image-1-768x848.png 768w" sizes="(max-width: 866px) 100vw, 866px" /></p>
<p>Strauss spells it out for us:</p>
<blockquote><p>Story Arc Literary Groups employs an approach common to many fake literary agency scams: promising to work on commission only, with no other fees due (note especially paragraph 5, which helpfully explains that “a reputable literary agent should not charge upfront fees”).  The aim of such solicitations, however, is always money, and writers who sign up with Story Arc soon discover this. In order for Story Arc to successfully pitch a book to traditional publishers, authors are told they must first “re-license” their book (a requirement that, as I’ve explained in another blog post, is completely fictional). As is typical for this type of scam, they’re referred to a “trusted” company to perform the service–in this case, an outfit called CreativeIP. The price tag: $5,000.</p></blockquote>
<p>Ouch!</p>
<p>Here&#8217;s another:</p>
<blockquote><p>Typical of fake literary agency scams, Zenith Literary is an aggressive solicitor. One writer who responded to this solicitation was told that in order to snag a traditional publisher’s interest, they needed to gather various “action items”, including “ten editorial reviews and endorsements” (hint: reviews and endorsements are nice, but they are absolutely not required by traditional publishers). To obtain these, the writer was referred to Verse Bound Solutions, a company with no apparent existence beyond a Wyoming business registration but active enough to phone the author and offer them ten book reviews for $3,000.</p></blockquote>
<p>And another:</p>
<p>The author who was targeted with [a solicitation from &#8220;ImplicitPress Literary Agency&#8221;] was asked to supply a variety of necessary “documents”; note #5, which is what this scam is hoping to sell (no publisher requires or cares about a book trailer):</p>
<p><img decoding="async" src="https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/image-2.png" alt="" width="450" /></p>
<p>If you&#8217;re itching for more such stories, <a href="https://writerbeware.blog/category/literary-agent-scams/">just go here</a>:</p>
<p><img loading="lazy" decoding="async" src="https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-09-07-at-14.12.47-1024x979.png" alt="" width="584" height="558" class="alignnone size-large wp-image-51013" srcset="https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-09-07-at-14.12.47-1024x979.png 1024w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-09-07-at-14.12.47-300x287.png 300w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-09-07-at-14.12.47-768x734.png 768w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-09-07-at-14.12.47-1536x1468.png 1536w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-09-07-at-14.12.47-2048x1958.png 2048w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-09-07-at-14.12.47-314x300.png 314w" sizes="(max-width: 584px) 100vw, 584px" /></p>
<p>What a world we live in.</p>
<p><strong>P.S.</strong>  In case you&#8217;re wondering about the title of this post, <a href="https://statmodeling.stat.columbia.edu/2010/05/12/alert_incompete/">see here for the relevant background</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/06/literary-agency-scams/feed/</wfw:commentRss>
			<slash:comments>10</slash:comments>
		
		
			</item>
		<item>
		<title>&#8220;I would have had a simple piece of advice: Say nothing.&#8221;</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/05/i-would-have-had-a-simple-piece-of-advice-say-nothing/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/05/i-would-have-had-a-simple-piece-of-advice-say-nothing/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Sun, 05 Jan 2025 14:26:36 +0000</pubDate>
				<category><![CDATA[Literature]]></category>
		<category><![CDATA[Sports]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51018</guid>

					<description><![CDATA[Amusing story here from Jonathan Bailey on &#8220;NaNoWriMo’s Massive AI Blunder.&#8221; I&#8217;m reminded of those people who lie about their running times in zero-stakes races. Bailey&#8217;s article is a pleasure to read because it&#8217;s just straight-up common sense.]]></description>
										<content:encoded><![CDATA[<p><a href="https://www.plagiarismtoday.com/2024/09/04/nanowrimos-massive-ai-blunder/">Amusing story here</a> from Jonathan Bailey on &#8220;NaNoWriMo’s Massive AI Blunder.&#8221;  I&#8217;m reminded of those people who lie about their running times in <a href="https://statmodeling.stat.columbia.edu/2012/11/07/that-last-satisfaction-at-the-end-of-the-career/">zero-stakes races</a>.</p>
<p>Bailey&#8217;s article is a pleasure to read because it&#8217;s just straight-up common sense.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/05/i-would-have-had-a-simple-piece-of-advice-say-nothing/feed/</wfw:commentRss>
			<slash:comments>2</slash:comments>
		
		
			</item>
		<item>
		<title>What are my goals?  What are their goals?  (How to prepare for that meeting.)</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/04/what-are-my-goals-what-are-their-goals-how-to-prepare-for-that-meeting/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/04/what-are-my-goals-what-are-their-goals-how-to-prepare-for-that-meeting/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Sat, 04 Jan 2025 14:50:27 +0000</pubDate>
				<category><![CDATA[Decision Analysis]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=50565</guid>

					<description><![CDATA[Corresponding with someone who had a difficult meeting coming up, where she was not sure how much to trust the person she was meeting with, I gave the following advice: Proceed under the assumption that they want to do things &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/04/what-are-my-goals-what-are-their-goals-how-to-prepare-for-that-meeting/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Corresponding with someone who had a difficult meeting coming up, where she was not sure how much to trust the person she was meeting with, I gave the following advice:</p>
<blockquote><p>Proceed under the assumption that they want to do things right.  I say this because if they&#8217;re gonna be defensive, then it doesn&#8217;t matter what you say; it&#8217;s not like you&#8217;re gonna sweet-talk them into opening up.  But if they do want to do better, then maybe there is some hope.</p></blockquote>
<p>My correspondent responded that the person she was meeting hadn&#8217;t been helpful up to this point:  &#8220;I always assume (and hope for) good intentions and a desire to do better. But I’ll admit I’m feeling less positive after a few days of not getting an answer.&#8221;</p>
<p>I continued:</p>
<blockquote><p>Many of these sorts of meetings require negotiation, and good negotiation often involves withholding of information or outright deception, and I&#8217;m not good at either of these things, so I don&#8217;t even try.  Instead I try some of the classic &#8220;Getting to Yes&#8221; strategies:<br />
(1) Before the meeting, I ask myself what are my goals:  my short-term goals for the meeting and my medium and long-term goals that I&#8217;m aiming for.<br />
(2) During the meeting, I explicitly ask the other parties what their goals are.</p>
<p>When I think of various counterproductive interactions I&#8217;ve had in the past, often it seems this has come in part because I was not clear on my goals or on the goals of the other parties; as a result we butted heads when we could&#8217;ve found a mutually-beneficial solution.  I&#8217;m including here some interactions with bad actors:  liars, cheats, etc.  Even when working with people you can&#8217;t trust, the general principles can apply.</p>
<p>It does not always make sense to tell the other parties what your goals are!  But, don&#8217;t worry, most people won&#8217;t ever ask, as they will typically be focused on trying to stand firm on some micro-issue or another.  Kinda like how amateur poker players are notorious for looking over and over again at their own hole cards and not looking enough at you.</p>
<p>The above advice may seem silly because you&#8217;re not involved in a negotiation at all!  Even so, if you have a sense of what your goals are and what their goals are, this could be helpful.  And be careful to distinguish goals from decision options.  A goal is &#8220;I would like X to happen&#8221;; a decision option is &#8220;I will do Y.&#8221;  It&#8217;s natural to think in terms of decision options, but I think this is limiting, compared to thinking about goals.</p>
<p>Anyway, that&#8217;s just my take from a mixture of personal experience and reading on decision making; I&#8217;ve done no direct research on the topic.</p></blockquote>
<p>The above techniques are not any sort of magic; they&#8217;re just an attempt to focus on what is important.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/04/what-are-my-goals-what-are-their-goals-how-to-prepare-for-that-meeting/feed/</wfw:commentRss>
			<slash:comments>10</slash:comments>
		
		
			</item>
		<item>
		<title>Echoing Eco:  From the logic of stories to posterior predictive simulation</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/03/echoing-eco-from-the-logic-of-stories-to-posterior-predictive-simulation/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/03/echoing-eco-from-the-logic-of-stories-to-posterior-predictive-simulation/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Fri, 03 Jan 2025 14:33:22 +0000</pubDate>
				<category><![CDATA[Bayesian Statistics]]></category>
		<category><![CDATA[Literature]]></category>
		<category><![CDATA[Miscellaneous Statistics]]></category>
		<category><![CDATA[Statistical Computing]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51008</guid>

					<description><![CDATA[&#8220;When I put Jorge in the library I did not yet know he was the murderer. He acted on his own, so to speak. And it must not be thought that this is an &#8216;idealistic&#8217; position, as if I were &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/03/echoing-eco-from-the-logic-of-stories-to-posterior-predictive-simulation/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><em>&#8220;When I put Jorge in the library I did not yet know he was the murderer.  He acted on his own, so to speak.  And it must not be thought that this is an &#8216;idealistic&#8217; position, as if I were saying that the characters have an autonomous life and the author, in a kind of trance, makes them behave as they themselves direct him.  That kind of nonsense belongs in term papers.  The fact is that the characters are obliged to act according to the laws of the world in which they live.  In other words, the narrator is the prisoner of his own premises.&#8221; &#8212; Umberto Eco, Postscript to The Name of the Rose (translated by William Weaver)</em></p>
<p>Perfectly put.  As I wrote less poetically <a href="https://statmodeling.stat.columbia.edu/2017/02/10/storytelling-predictive-model-checking/">a few years ago</a>, the development of a story is a working-out of possibilities, and that’s why it makes sense that authors can be surprised at how their own stories come out.  In statistics jargon, the surprise we see in a story is a form of predictive check, a recognition that a scenario, if carried out logically, can lead to unexpected places.</p>
<p><a href="http://stat.columbia.edu/~gelman/research/published/isr.pdf">In statistics</a>, one reason we make predictions is to do <a href="http://stat.columbia.edu/~gelman/research/published/A6n41.pdf">predictive checks</a>, to elucidate the implications of a model, in particular what it says (probabilistically) regarding observable outcomes, which can then be checked with existing or new data.</p>
<p>To put it in <a href="http://stat.columbia.edu/~gelman/research/published/storytelling.pdf">storytelling terms</a>, if you tell a story and it leads to a nonsensical conclusion, this implies there’s something wrong with your narrative logic or with your initial scenario.</p>
<p>Again, I really like how Eco frames the problem, reconciling the agency of the author (who is the one who comes up with premise and the rules of the game and who works out their implications) and the apparent autonomy of the character (which is a consequence of the logic of the story).</p>
<p>This also connects to <a href="https://statmodeling.stat.columbia.edu/2023/11/18/i-disagree-with-geoff-hinton-regarding-glorified-autocomplete/">a discussion we had a year ago</a> about chatbots.  As I wrote at the time, a lot of what I do at work&#8212;or when blogging!&#8212;is a sort of autocomplete, where I start with some idea and work out its implications.  Indeed, an important part of the writing process is to get into a flow state where the words, sentences, and paragraphs come out smoothly, and in that sense there&#8217;s no other way to do this than with some sort of autocomplete.  Autocomplete isn&#8217;t everything&#8212;sometimes I need to stop and think, make plans, do some math&#8212;but it&#8217;s a lot.</p>
<p>Different people do autocomplete in different ways.  Just restricting ourselves to bloggers here, give the same prompt to Jessica, Bob, Phil, and Lizzie, and you&#8217;ll get four much different posts&#8212;and, similarly, Umberto Eco&#8217;s working out of the logic of a murder in a medieval monastery will come out different from yours.  Not even considering the confounding factor that we get to choose the &#8220;prompts&#8221; for our blog posts, and Eco picked his scenario because he thought it would be fruitful ground for a philosophical novel.  Writing has value, in the same way that prior or posterior simulation has value:  we don&#8217;t know how things will come out any more than we can know the millionth digit of pi without doing the damn calculation.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/03/echoing-eco-from-the-logic-of-stories-to-posterior-predictive-simulation/feed/</wfw:commentRss>
			<slash:comments>11</slash:comments>
		
		
			</item>
		<item>
		<title>Progress in 2024 (Jessica)</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/02/progress-in-2024-jessica/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/02/progress-in-2024-jessica/#comments</comments>
		
		<dc:creator><![CDATA[Jessica Hullman]]></dc:creator>
		<pubDate>Thu, 02 Jan 2025 18:08:37 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Miscellaneous Statistics]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51526</guid>

					<description><![CDATA[2024 was an enjoyable year. Below are a few things I did. Conference and journal papers published Improving out-of-population prediction: The complementary effects of model assistance and judgmental bootstrapping. International Journal of Forecasting. Hardy, M. D., Zhang, S., Hullman, J., &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/02/progress-in-2024-jessica/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><span style="font-weight: 400">2024 was an enjoyable year. Below are a few things I did.</span></p>
<p><strong>Conference and journal papers published</strong></p>
<ul>
<li style="font-weight: 400"><a href="https://www.semanticscholar.org/paper/Improving-out-of-population-prediction%3A-The-effects-Hardy-Zhang/0f7903289a7cd6986df75897f6fd126e4c3a517d"><span style="font-weight: 400">Improving out-of-population prediction: The complementary effects of model assistance and judgmental bootstrapping</span></a><span style="font-weight: 400">. International Journal of Forecasting. Hardy, M. D., Zhang, S., Hullman, J., Hofman, J. M., and Goldstein, D. G. </span></li>
<li style="font-weight: 400"><a href="https://journals.sagepub.com/doi/abs/10.1177/23727322241278687"><span style="font-weight: 400">What to Consider When Considering Differential Privacy for Policy</span></a><span style="font-weight: 400">. Policy Insights from the Behavioral and Brain Sciences (PIBBS). Nanayakkara, P. and Hullman, J.</span></li>
<li style="font-weight: 400"><a href="https://dl.acm.org/doi/10.1109/TVCG.2024.3456402"><span style="font-weight: 400">VMC: A Grammar for Visualizing Statistical Model Checks</span></a><span style="font-weight: 400">. IEEE Transactions of Visualization &amp; Computer Graphics (Proceedings of IEEE VIS 2024). Guo, Z., Kale, A., Kay, M., and Hullman, J.</span></li>
<li style="font-weight: 400"><a href="https://www.science.org/doi/10.1126/sciadv.adk3452"><span style="font-weight: 400">REFORMS: Consensus-based Recommendations for Machine-learning-based Science</span></a><span style="font-weight: 400">. Science Advances, 10 (18). Kapoor, S., Cantrell, E. M., Peng, K., Pham, T. H., Bail, C., Gundersen, O. E., Hofman, J., Hullman, J., Lones, M., Malik, M., Nanayakkara, P., Poldrack, R., Raji, I. D., Roberts, M., Salganik, M., Serra-Garcia, M., Stewart, B., Vandewiele, G., and Narayanan, A.</span></li>
<li style="font-weight: 400"><a href="https://ojs.aaai.org/index.php/AIES/article/view/31656/33823"><span style="font-weight: 400">A Conceptual Framework for Ethical Evaluation of Machine Learning Systems</span></a><span style="font-weight: 400">. Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society. Gupta, N. R., Hullman, J., and Subramonyam, H.</span></li>
<li style="font-weight: 400"><a href="https://www.computer.org/csdl/proceedings-article/sp/2024/313000a231/1WPcYT0BBmw"><span style="font-weight: 400">Measure-Observe-Remeasure: An Interactive Paradigm for Differentially-Private Exploratory Analysis</span></a><span style="font-weight: 400">. IEEE symposium on Privacy &amp; Security (Proceedings of S&amp;P 2024). Nanayakkara, P., Kim, H., Wu, Y., Sarvghad, A., Mahyar, N., Miklau, G., and Hullman, J. </span></li>
<li style="font-weight: 400"><a href="https://dl.acm.org/doi/fullHtml/10.1145/3630106.3658901"><span style="font-weight: 400">A Decision Theoretic Framework for Measuring AI Reliance</span></a><span style="font-weight: 400">. ACM Conference on Fairness, Accountability, &amp; Transparency (Proceedings of FAccT 2024). Guo, Z., Wu, Y., Hartline, J., and Hullman, J.</span></li>
<li style="font-weight: 400"><a href="https://dl.acm.org/doi/10.1145/3613904.3642446"><span style="font-weight: 400">Evaluating the Utility of Conformal Prediction Sets for AI-Advised Image Labeling</span></a><span style="font-weight: 400">. ACM Conference on Human Factors in Computing Systems (Proceedings of CHI 2024). Zhang, D., Chatzimparmpas, A., Kamali, N., and Hullman, J.</span></li>
<li style="font-weight: 400"><a href="https://dl.acm.org/doi/10.1145/3613904.3642442"><span style="font-weight: 400">Erie: A Declarative Grammar for Data Sonification</span></a><span style="font-weight: 400">. ACM Conference on Human Factors in Computing Systems (Proceedings of CHI 2024. Kim, H., Kim, Y-S., and Hullman, J.</span></li>
<li style="font-weight: 400"><a href="https://dl.acm.org/doi/10.1145/3613904.3642375"><span style="font-weight: 400">Milliways: Taming Multiverses through Principled Evaluation of Data Analysis Paths</span></a><span style="font-weight: 400">. ACM Conference on Human Factors in Computing Systems (Proceedings of CHI 2024. Sarma, A., Hwang, K., Hullman, J., and Kay, M.</span></li>
<li style="font-weight: 400"><a href="http://stat.columbia.edu/~gelman/research/published/causal_quartets_tas.pdf"><span style="font-weight: 400">Causal quartets: Different ways to attain the same average treatment effect. </span></a><span style="font-weight: 400">American Statistician 78</span><a href="http://stat.columbia.edu/~gelman/research/published/causal_quartets_tas.pdf"><span style="font-weight: 400">.</span></a><span style="font-weight: 400"> Gelman, A., Hullman, J., and Kennedy, L.</span></li>
</ul>
<p><span style="font-weight: 400">Three of my collaborators above were Ph.D. students I advised, who graduated in 2024! Congrats to <a href="https://priyakalot.github.io/">Priyanka Nanayakkara</a> (now a postdoc at Harvard CRCS), <a href="https://hyeok.me/">Hyeok Kim</a> (now a postdoc at University of Washington CS, on the academic job market), and <a href="https://www.dpzhang.com/">Dongping Zhang</a> (now a research scientist at NREL). </span></p>
<p><strong>Talks (that are available online)</strong></p>
<ul>
<li style="font-weight: 400"><a href="https://www.youtube.com/watch?v=GApHJudc4G4"><span style="font-weight: 400">Benchmarking Visualization for Decision-Making</span></a><span style="font-weight: 400">. MIT Code Conference, Oct. 2024.</span></li>
<li style="font-weight: 400"><a href="https://www.youtube.com/watch?v=EJR71K4hnYI&amp;list=PLuD_SqLtxSdWfDA_GJLFqcj1IcU3ujkBQ&amp;index=3"><span style="font-weight: 400">Data analysis and imagination</span></a><span style="font-weight: 400">. </span><span style="font-weight: 400">Alan Turing Institute, June 2024.</span></li>
</ul>
<p><span style="font-weight: 400">I gave a several other talks that I think exist somewhere online, but I can’t find the links. </span></p>
<p><strong>Workshop participation/organization </strong></p>
<p><span style="font-weight: 400">The highlights of the year for me were workshops I attended over the summer. Getting to travel to interesting places to think deeply about topics you find fascinating is truly a privilege. </span></p>
<ul>
<li style="font-weight: 400"><a href="https://www.birs.ca/events/2024/5-day-workshops/24w5283"><span style="font-weight: 400">Bridging Prediction and Intervention Problems in Social Systems</span></a><span style="font-weight: 400">. Organized by Lydia Liu, Inioluwa Deborah Raji, Angela Zhou, and Arvind Narayanan. Banff Canada, June 2024. </span></li>
<li style="font-weight: 400"><a href="https://www.turing.ac.uk/events/tmcf-workshop-navigating-garden-forking-paths"><span style="font-weight: 400">Navigating the garden of forking paths/Theoretical foundations for interactive data analysis in data-driven science</span></a><span style="font-weight: 400">. Organized by Cagatay Turkay and Roger Beecham. London, June 2024.</span></li>
<li style="font-weight: 400"><a href="https://people.eecs.berkeley.edu/~brecht/n_of_1/"><span style="font-weight: 400">Workshop on Individualized Decision-Making</span></a><span style="font-weight: 400">. Organized by Ben Recht. Berkeley, CA, July 2024.</span></li>
<li style="font-weight: 400"><span style="font-weight: 400">Apple Workshop on Human-Centered Machine Learning. Organized by Kareem Bedri, Leah Findlater, Dominik Moritz, and Jeff Nichols. Cupertino, CA, Aug 2024.</span></li>
</ul>
<p><span style="font-weight: 400">I co-organized a few other workshops I enjoyed:</span></p>
<ul>
<li style="font-weight: 400"><a href="https://www.ideal-institute.org/2024/08/21/workshop-on-theoretical-foundations-of-human-ai-complementarity/"><span style="font-weight: 400">Theoretical Foundations of Human-AI Complementarity</span></a><span style="font-weight: 400">. With Jason Hartline. Northwestern University, Sept. 2024.</span></li>
<li style="font-weight: 400"><a href="https://sites.google.com/berkeley.edu/bb-stat/home"><span style="font-weight: 400">Statistical Frontiers in LLMs and Foundation Models</span></a><span style="font-weight: 400"><span style="font-weight: 400">. With Anastasios Angelopoulos, Stephen Bates, Alex D&#8217;Amour, Fanny Yang, Sophia Sun, and Tatsunori Hashimoto. NeurIPS, Vancouver, Dec. 2024.</span></span></li>
</ul>
<p><span style="font-weight: 400">My goal for next year is to do more creative writing. I would be thrilled if I could write even a couple poems I’m happy with. </span></p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/02/progress-in-2024-jessica/feed/</wfw:commentRss>
			<slash:comments>2</slash:comments>
		
		
			</item>
		<item>
		<title>Advice for weighting the results of conjoint analyses/experiments</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/02/advice-for-weighting-the-results-of-conjoint-analyses-experiments/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/02/advice-for-weighting-the-results-of-conjoint-analyses-experiments/#respond</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Thu, 02 Jan 2025 14:22:42 +0000</pubDate>
				<category><![CDATA[Causal Inference]]></category>
		<category><![CDATA[Miscellaneous Statistics]]></category>
		<category><![CDATA[Political Science]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51005</guid>

					<description><![CDATA[Someone asked me, &#8220;What advice do you have for weighting the results of conjoint analyses/experiments?” I replied that a conjoint experiment is basically a regression analysis. Here we discuss survey weights and regression, and here&#8217;s the relevant bit: So the &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/02/advice-for-weighting-the-results-of-conjoint-analyses-experiments/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Someone asked me, &#8220;What advice do you have for weighting the results of conjoint analyses/experiments?”</p>
<p>I replied that a conjoint experiment is basically a regression analysis.  <a href="http://stat.columbia.edu/~gelman/research/unpublished/weight_regression.pdf">Here we discuss</a> survey weights and regression, and here&#8217;s the relevant bit:</p>
<p><img loading="lazy" decoding="async" src="https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-08-29-at-20.56.02-1024x906.png" alt="" width="584" height="517" class="alignnone size-large wp-image-51006" srcset="https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-08-29-at-20.56.02-1024x906.png 1024w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-08-29-at-20.56.02-300x265.png 300w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-08-29-at-20.56.02-768x679.png 768w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-08-29-at-20.56.02-1536x1358.png 1536w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-08-29-at-20.56.02-339x300.png 339w, https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/09/Screenshot-2024-08-29-at-20.56.02.png 1920w" sizes="(max-width: 584px) 100vw, 584px" /></p>
<p>So the first paper I recommend is Winship, C., and Radbill, L. (1994). Sampling weights and regression analysis. Sociological Methods and Research 23, 230–257.</p>
<p>I also recommend <a href="https://www.cambridge.org/core/journals/political-analysis/article/improving-the-external-validity-of-conjoint-analysis-the-essential-role-of-profile-distribution/B911EF14513292A24ECB4AC4BAA3FA6B">this recent paper</a> by Brandon de la Cuesta, Naoki Egami, Kosuke Imai, Improving the External Validity of Conjoint Analysis: The Essential Role of Profile Distribution, which addresses issues of average predictive comparisons that have come up many times before in this space, for example <a href="https://statmodeling.stat.columbia.edu/2019/12/29/do-we-still-recommend-average-predictive-comparisons-click-here-to-find-the-surprising-answer/">here</a> and <a href="https://statmodeling.stat.columbia.edu/2022/07/12/clearing-up-confusions-about-the-interpretation-of-logistic-regression-coefficients-on-the-probability-scale/">here</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/02/advice-for-weighting-the-results-of-conjoint-analyses-experiments/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Newly published in 2024</title>
		<link>https://statmodeling.stat.columbia.edu/2025/01/01/published-or-accepted-for-publication-in-2024/</link>
					<comments>https://statmodeling.stat.columbia.edu/2025/01/01/published-or-accepted-for-publication-in-2024/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Wed, 01 Jan 2025 14:16:26 +0000</pubDate>
				<category><![CDATA[Decision Analysis]]></category>
		<category><![CDATA[Literature]]></category>
		<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Miscellaneous Statistics]]></category>
		<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51519</guid>

					<description><![CDATA[Our big item is the book Active Statistics: Stories, Games, Problems, and Hands-on Demonstrations for Applied Regression and Causal Inference (Andrew Gelman and Aki Vehtari). Then there are the recently published research articles: [2025] Hierarchical Bayesian models to mitigate systematic &#8230; <a href="https://statmodeling.stat.columbia.edu/2025/01/01/published-or-accepted-for-publication-in-2024/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Our big item is the book <a href="https://stat.columbia.edu/~gelman/active-statistics/">Active Statistics: Stories, Games, Problems, and Hands-on Demonstrations for Applied Regression and Causal Inference</a> (Andrew Gelman and Aki Vehtari).</p>
<p>Then there are the recently published <a href="http://stat.columbia.edu/~gelman/research/published/">research articles</a>:</p>
<ul>
<li><a href="http://stat.columbia.edu/~gelman/research/published/MeasurementModels.pdf">[2025] Hierarchical Bayesian models to mitigate systematic disparities in prediction with proxy outcomes. Journal of the Royal Statistical Society A.</a>(Jonas Mikhaeil, Andrew Gelman, and Philip Greengard)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/piranha_ams_notices.pdf">[2025] The piranha problem: Large effects swimming in a small pond. Notices of the American Mathematical Society.</a> (Christopher Tosh, Philip Greengard, Ben Goodrich, Andrew Gelman, Aki Vehtari, sand Daniel Hsu)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/2311.02726.pdf">[2025] For how many iterations should we run Markov chain Monte Carlo? In Handbook of Markov Chain Monte Carlo, second edition.</a> (Charles C. Margossian and Andrew Gelman)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/election_forecasting.pdf">[2024] Why forecast an election that’s too close to call? Nature 634, 1019.</a> (Andrew Gelman)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/2024_Election_Forecasting_Review.pdf">[2024] Grappling with uncertainty in forecasting the 2024 U.S. presidential election. Harvard Data Science Review 6 (4).</a> (Andrew Gelman, Ben Goodrich, and Geonhee Han)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/noise_review.pdf">[2024] Review of &#8220;Noise: A Flaw in Human Judgment,&#8221; by Daniel Kahneman, Olivier Sibony, and Cass R. Sunstein. Chance 37 (3), 70-72.</a>(Gaurav Sood and Andrew Gelman)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/healing3.pdf">[2024] How statistical challenges and misreadings of the literature combine to produce unreplicable science: An example from psychology. Advances in Methods and Practices in Psychological Science.</a> (Andrew Gelman and Nicholas J. L. Brown)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/amalgamating7.pdf">[2024] Statistics as a social activity: Attitudes toward amalgamating evidence. Entropy 26 (8), 652.</a> (Andrew Gelman and Keith O’Rourke)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/nestedRhat.pdf">[2024] Nested R-hat: Assessing the convergence of Markov chain Monte Carlo when running many short chains. Bayesian Analysis.</a> (Charles C. Margossian, Matthew D. Hoffman, Pavel Sountsov, Lionel Riou-Durand, Aki Vehtari, and Andrew Gelman)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/2209.01773.pdf">[2024] Using leave-one-out cross-validation (LOO) in a multilevel regression and poststratification (MRP) workflow: A cautionary tale. Statistics in Medicine 43, 953-982.</a> (Swen Kuh, Lauren Kennedy, Qixuan Chen, and Andrew Gelman)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/donoho_hdsr.pdf">[2024] Hopes and limitations of reproducible statistics and machine learning. Harvard Data Science Review 6 (1).</a> (Andrew Gelman)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/psis6_jmlr.pdf">[2024] Pareto smoothed importance sampling. Journal of Machine Learning Research 25 (72).</a> (Aki Vehtari, Daniel Simpson, Andrew Gelman, Yuling Yao, and Jonah Gabry)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/pval_RCTs_rev3.pdf">[2024] A new look at p-values for randomized clinical trials. NEJM Evidence 3 (1).</a> (Erik van Zwet, Andrew Gelman, Sander Greenland, Guido Imbens, Simon Schwab, and Steven N. Goodman)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/2211.02383.pdf">[2024] Simulation-based calibration checking for Bayesian computation: The choice of test quantities shapes sensitivity. Bayesian Analysis.</a> (Martin Modrák, Angie H. Moon, Shinyoung Kim, Paul Bürkner, Niko Huurre, Kateřina Faltejsková, Andrew Gelman, and Aki Vehtari)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/jcp.pdf">[2024] Before data analysis: Additional recommendations for designing experiments to learn about the world. Journal of Consumer Psychology 34, 190-191.</a> (Andrew Gelman)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/Research_Build-Your-Own-Statistics%20Course.pdf">[2024] In pursuit of campus-wide data literacy: A guide to developing a statistics course for students in non-quantitative fields. Journal of Statistics and Data Science Education 32, 241-252.</a> (Alexis Lerner and Andrew Gelman)</li>
<li><a href="http://stat.columbia.edu/~gelman/research/published/causal_quartets_tas.pdf">[2024] Causal quartets: Different ways to attain the same average treatment effect. American Statistician 78, 267-272.</a> (Andrew Gelman, Jessica Hullman, and Lauren Kennedy)</li>
</ul>
<p>It&#8217;s a good mix:  some technical work, some research methods, some applications, some teaching material, some reviews. Lots of collaborators!</p>
<p>If you want to see what&#8217;s coming next, you can check out the lists of <a href="http://stat.columbia.edu/~gelman/research/unpublished/">unpublished</a> and <a href="http://stat.columbia.edu/~gelman/research/unwritten/">unwritten</a> articles.</p>
<p>Here are our audios and videos from the past year:</p>
<ul>
<li><a href="https://www.youtube.com/watch?v=5BNVBiPpYlY">Bayesian Workflow</a> (Conversation with Charles Margossian for the MetrumRG podcast, 22 Oct 2024)</li>
<li><a href="https://www.youtube.com/watch?v=EqNbiNcQuKo">Tough Choices in Election Forecasting: All the Things That Can Go Wrong</a> (Presented at the Washington Statistical Society, 11 Oct 2024)</li>
<li><a href="https://www.youtube.com/watch?v=7ZYTWrh8YgI">The Political Content of Unreplicable Research</a> (Presented at the Stanford Classical Liberalism Seminar, 3 Oct 2024)</li>
<li><a href="https://www.youtube.com/watch?v=okv48JwsBw4">Fooling Yourself Less: The Art of Statistical Thinking in AI</a> (Conversation for the High Signal podcast, 18 Sep 2024)</li>
<li><a href="https://www.youtube.com/watch?v=eVNiG7AzPwY">Holes in Bayesian Statistics</a> (Presented at the International Society for Bayesian Analysis meeting, 3 Jul 2024)</li>
<li><a href="https://www.youtube.com/watch?v=Y-UySxjiTks&amp;list=PLuD_SqLtxSdWfDA_GJLFqcj1IcU3ujkBQ&amp;index=2">Beyond the Black Box: Toward a New Paradigm of Statistics in Science</a> (Presented at the Alan Turing Institute, 20 Jun 2024)</li>
</ul>
<p>Last but not least, our blog posts of 2024. We had 540 posts with 11,483 total comments:</p>
<p>It’s bezzle time: The Dean of Engineering at the University of Nevada gets paid $372,127 a year and wrote a paper that’s so bad, you can’t believe it. (204 comments)<br />
In some cases academic misconduct doesn’t deserve a public apology (177 comments)<br />
Getting a pass on evaluating ways to improve science (154 comments)<br />
Niall Ferguson, J. D. Vance, George Washington, and Jesus (129 comments)<br />
Stabbers gonna stab — fraud edition (124 comments)<br />
Suspicious data pattern in recent Venezuelan election (111 comments)<br />
This well-known paradox of R-squared is still buggin me. Can you help me out? (105 comments)<br />
Reflections on the recent election (103 comments)<br />
The mainstream press is failing America (UK edition) (99 comments)<br />
The Behavioural Insights Team decided to scare people. (98 comments)<br />
If you want to play women’s tennis at the top level, there’s a huge benefit to being ____. Not just ____, but exceptionally ___, outlier-outlier ___. (And what we can learn about social science from this stylized fact.) (94 comments)<br />
“Is it really ‘the economy, stupid’?” (91 comments)<br />
Bad stuff going down at the American Sociological Association (84 comments)<br />
How to think about the claim by Justin Wolfers that “the income of the average American will double approximately every 39 years”? (84 comments)<br />
Polling averages and political forecasts and what do you really think is gonna happen in November? (83 comments)<br />
Intelligence is whatever machines cannot (yet) do (82 comments)<br />
On the border between credulity and postmodernism: The case of the UFO’s-as-space-aliens media insiders (78 comments)<br />
Understanding p-values: Different interpretations can be thought of not as different “philosophies” but as different forms of averaging. (78 comments)<br />
“Not once in the twentieth century . . . has a single politician, actor, athlete, or surgeon emerged as a first-rate novelist, despite the dismayingly huge breadth of experience each profession affords.” (77 comments)<br />
Abraham Lincoln and confidence intervals (77 comments)<br />
What is the prevalence of bad social science? (76 comments)<br />
Whassup with those economists who predicted a recession that then didn’t happen? (74 comments)<br />
Feedback on the blog—this is your chance! (74 comments)<br />
If school funding doesn’t really matter, why do people want their kid’s school to be well funded? (69 comments)<br />
“It’s a very short jump from believing kale smoothies are a cure for cancer to denying the Holocaust happened.” (69 comments)<br />
On lying politicians and bullshitting scientists (67 comments)<br />
My comments on Nate Silver’s comments on the Fivethirtyeight election forecast (66 comments)<br />
My suggestion for the 2028 Olympics (66 comments)<br />
The statistical controversy over “White Rural Rage: the Threat to American Democracy” (and a comment about post-publication review) (65 comments)<br />
HMC fails when you initialize at the mode (63 comments)<br />
Selection bias leads to confusion about the relative stability of deterministic and stochastic algorithms (63 comments)<br />
Wendy Brown: “Just as nothing is more corrosive to serious intellectual work than being governed by a political programme (whether that of states, corporations, or a revolutionary movement), nothing is more inapt to a political campaign than the unending reflexivity, critique and self-correction required of scholarly inquiry.” (63 comments)<br />
Bayesians are frequentists. (62 comments)<br />
More red meat for you AI skeptics out there (61 comments)<br />
The River, the Village, and the Fort: Nate Silver’s new book, “On the Edge” (61 comments)<br />
“Zombie Ideas” in psychology, from personality profiling to lucky golf balls (61 comments)<br />
Andrew Gelman is not the science police because there is no such thing as the science police (60 comments)<br />
With journals, it’s all about the wedding, never about the marriage. (59 comments)<br />
Stanford medical school professor misrepresents what I wrote (but I kind of understand where he’s coming from) (59 comments)<br />
Hey—let’s collect all the stupid things that researchers say in order to deflect legitimate criticism (58 comments)<br />
Bayesian statistics: the three cultures (58 comments)<br />
A quick simulation to demonstrate the wild variability of p-values (58 comments)<br />
How to think about the effect of the economy on political attitudes and behavior? (57 comments)<br />
Forking paths in LLMs for data analysis (57 comments)<br />
Two kings, a royal, a knight, and three princesses walk into a bar (Nobel prize edition) (57 comments)<br />
Prediction markets in 2024 and poll aggregation in 2008 (57 comments)<br />
The New York Young Republican Club (56 comments)<br />
Holes in Bayesian statistics (my talk tomorrow at the Bayesian conference, based on work with Yuling) (56 comments)<br />
Abortion crime controversy update (55 comments)<br />
Some references and discussions on the foundations of probability—not the math so much as its connection to the real world, including the claim that “Pr(aliens exist on Neptune that can rap battle) = .137” (54 comments)<br />
Props to the liberal anticommunists of the 1930s-1950s (54 comments)<br />
Interpreting recent Iowa election poll using a rough Bayesian partition of error (54 comments)<br />
How would the election turn out if Biden or Trump were replaced by a different candidate? (53 comments)<br />
Uncertainty in games: How to get that balance so that there’s a motivation to play well, but you can still have a chance to come back from behind? (52 comments)<br />
Decisions of parties to run moderate or extreme candidates (52 comments)<br />
Benefit of Stanford: Are there connections between unethical behavior in science promotion and cheating in private life? (51 comments)<br />
When all else fails, add a code comment (50 comments)<br />
The feel-good open science story versus the preregistration (who do you think wins?) (49 comments)<br />
Deadwood (49 comments)<br />
Why are all these school cheating scandals happening? (48 comments)<br />
Is marriage associated with happiness for men or for women? Or both? Or neither? (48 comments)<br />
He took public funds and falsified his data. Are they gonna make him pay back the $19 million? (48 comments)<br />
Arnold Foundation and Vera Institute argue about a study of the effectiveness of college education programs in prison. (46 comments)<br />
How do you interpret standard errors from a regression fit to the entire population? (46 comments)<br />
It’s Harvard time, baby: “Kerfuffle” is what you call it when you completely botched your data but you don’t want to change your conclusions. (46 comments)<br />
Where have all the count words gone? In defense of “fewer” and “among” (45 comments)<br />
Arguing about bitcoin (44 comments)<br />
Polling by asking people about their neighbors: When does this work? Should people be doing more of it? And the connection to that French dude who bet on Trump (44 comments)<br />
Prediction isn’t everything, but everything is prediction (43 comments)<br />
Who wrote the music for In My Life? Three Bayesian analyses (43 comments)<br />
Honesty and transparency are not enough: politics edition (43 comments)<br />
What’s gonna happen between now and November 5? (42 comments)<br />
What’s the story behind that paper by the Center for Open Science team that just got retracted? (42 comments)<br />
Torment executioners in Reno, Nevada, keep tormenting us with their publications. (41 comments)<br />
“You want to gather data to determine which of two students is a better basketball shooter. You plan to have each student take N shots and then compare their shooting percentages. Roughly how large does N have to be for you to have a good chance of distinguishing a 30% shooter from a 40% shooter?” (41 comments)<br />
Why are we making probabilistic election forecasts? (and why don’t we put so much effort into them?) (41 comments)<br />
The appeal of New York Times columnist David Brooks . . . Yeah, I know this all sounds like a nutty “it’s wheels within wheels, man” sort of argument, but I’m serious here! (40 comments)<br />
Sympathy for the Nudgelords: Vermeule endorsing stupid and dangerous election-fraud claims and Levitt promoting climate change denial are like cool dudes in the 60s wearing Che T-shirts and thinking Chairman Mao was cool—we think they’re playing with fire, they think they’re cute contrarians pointing out contradictions in the system. For a certain kind of person, it’s fun to be a rogue. (40 comments)<br />
Freakonomics does it again (not in a good way). Jeez, these guys are credulous: (40 comments)<br />
Mister P and Stan go to Bangladesh . . . (39 comments)<br />
“Things are Getting So Politically Polarized We Can’t Measure How Politically Polarized Things are Getting” (39 comments)<br />
Opposition (38 comments)<br />
“My basic question is do we really need data to be analysed by both methods?” (38 comments)<br />
The most interesting part of the story is that the publisher went through all these steps of reviewing and revising. If they just want to make money by publishing crap, why bother engaging outside reviewers at all? (38 comments)<br />
A new argument for estimating the probability that your vote will be decisive (37 comments)<br />
Kamala Harris gets coveted xkcd endorsement. (37 comments)<br />
Implicitly denying the controversy associated with the Implicit Association Test. (Whassup with the American Association of Arts &amp; Sciences?) (37 comments)<br />
Simulation to understand two kinds of measurement error in regression (36 comments)<br />
I strongly doubt that any human has ever typed the phrase, “torment executioners,” on any keyboard—except, of course, in discussions such as this. (36 comments)<br />
What is the purpose of a methods section? (36 comments)<br />
The four principles of Barnard College: Respect, empathy, kindness . . . and censorship? (35 comments)<br />
Whooping cough! How to respond to fatally-flawed papers? An example, in a setting where the fatal flaw is subtle, involving a confounding of time and cohort effects (35 comments)<br />
The election is coming: What forecasts should we trust? (35 comments)<br />
Prediction markets and the need for “dumb money” as well as “smart money” (35 comments)<br />
“The Exceptions: Nancy Hopkins, MIT, and the Fight for Women in Science” (35 comments)<br />
Shreddergate! A fascinating investigation into possible dishonesty in a psychology experiment (34 comments)<br />
20 years of blogging . . . What have been your favorite posts? (34 comments)<br />
That’s what happens when you try to run the world while excluding 99.8% of the population (34 comments)<br />
The Theory and Practice of Oligarchical Collectivism (34 comments)<br />
What’s the problem, “math snobs” or rich dudes who take themselves too seriously and are enabled in that by the news media? (33 comments)<br />
(Trying to) clear up a misunderstanding about decision analysis and significance testing (33 comments)<br />
“Take a pass”: New contronym just dropped. (33 comments)<br />
Well, today we find our heroes flying along smoothly… (33 comments)<br />
“How a simple math error sparked a panic about black plastic kitchen utensils”: Does it matter when an estimate is off by a factor of 10? (33 comments)<br />
“Why do medical tests always have error rates?” (32 comments)<br />
Toward a unified theory of bad science and bad scholarship (32 comments)<br />
Michael Clayton in NYC (32 comments)<br />
I’ve been mistaken for a chatbot (31 comments)<br />
Here’s my excuse for using obsolete, sub-optimal, or inadequate statistical methods or using a method irresponsibly. (31 comments)<br />
“Participants reported being hungrier when they walked into the café (mean = 7.38, SD = 2.20) than when they walked out [mean = 1.53, SD = 2.70, F(1, 75) = 107.68, P &lt; 0.001].&#8221; (30 comments)<br />
“Don’t feed the trolls” and the troll semi-bluff (30 comments)<br />
Bad parenting in the news, also, yeah, lots of kids don’t believe in Santa Claus (30 comments)<br />
What do the data say about Kamala Harris’s electability? (30 comments)<br />
When is calibration enough? (30 comments)<br />
“I wonder just what it takes to get people to conclude that a research seam has been mined to the point of exhaustion.” (30 comments)<br />
Extinct Champagne grapes? I can be even more disappointed in the news media (29 comments)<br />
Relating t-statistics and the relative width of confidence intervals (29 comments)<br />
Report of average change from an Alzheimer’s drug: I don’t get the criticism here. (29 comments)<br />
Obnoxious receipt from Spirit Airlines (29 comments)<br />
The recent Iranian election: Should we be suspicious that the vote totals are all divisible by 3? (29 comments)<br />
Bill James hangs up his hat. Also some general thoughts about book writing vs. blogging. Also I push back against James’s claim about sabermetrics and statistics. (29 comments)<br />
Unsolicited feedback on your research from the LLMs at large (29 comments)<br />
Defining statistical models in JAX? (29 comments)<br />
She wants to know what are best practices on flagging bad responses and cleaning survey data and detecting bad responses. Any suggestions from the tidyverse or crunch.io? (29 comments)<br />
What to do with age? (including a regression predictor linearly and also in discrete steps) (28 comments)<br />
“Our troops with aching hearts were obliged to fire a part of the town as a punishment.” (28 comments)<br />
The Rider (28 comments)<br />
“Trivia question for you. I kept temperature records for 100 days one year in Boston, starting August 15th (day “0”). What would you guess is the correlation between day# and temp? r=???” (28 comments)<br />
Anti-immigration attitudes: they didn’t want a bunch of Hungarian refugees coming in the 1950s (28 comments)<br />
What genre of writing is AI-generated poetry? (28 comments)<br />
“I work in a biology lab . . . My PI proposed a statistical test that I think is nonsense. . .” (28 comments)<br />
Blog is adapted to laptops or desktops, not to smartphones or pads. (27 comments)<br />
“On the uses and abuses of regression models: a call for reform of statistical practice and teaching”: We’d appreciate your comments . . . (27 comments)<br />
Inspiring story from a chemistry classroom (27 comments)<br />
Which books, papers, and blogs are in the Bayesian canon? (27 comments)<br />
3M misconduct regarding knowledge of “forever chemicals”: As is so often the case, the problem was in open sight for a long time before anything was done (27 comments)<br />
God is in every leaf of every tree—comic book movies edition. (26 comments)<br />
Our new Substack newsletter: The Future of Statistical Modeling! (26 comments)<br />
Pinker was right, I was wrong. (26 comments)<br />
Does this study really show that lesbians and bisexual women die sooner than straight women? Disparities in Mortality by Sexual Orientation in a Large, Prospective JAMA Paper (26 comments)<br />
A cook, a housemaid, a gardener, a chauffeur, a nanny, a philosopher, and his wife . . . (26 comments)<br />
MCMC draws cannot fill the posterior in high dimensions (26 comments)<br />
Beyond junk science: How to go forward (26 comments)<br />
Instability of win probability in election forecasts (with a little bit of R) (26 comments)<br />
Mark Twain on chatbots (26 comments)<br />
Fake data on the honeybee waggle dance, followed by the inevitable “It is important to note that the conclusions of our studies remain firm and sound.” (26 comments)<br />
Make a hypothesis about what you expect to see, every step of the way. A manifesto: (26 comments)<br />
“Of course, this could conceivably be a case of near unbelievable luck: A flawed analysis based on wrong assumptions gave an unusually large causal effect estimate – but the misguided result just happened to be correct. We can imagine how the research team huddled nervously around the computer terminal biting their nails and silently praying as they executed their updated Stata code, only to erupt in joy and celebration as the results appeared on screen and revealed they were right all along. . . .” (26 comments)<br />
A very interesting discussion by Roy Sorensen of the interesting-number paradox (26 comments)<br />
I love this paper but it’s barely been noticed. (25 comments)<br />
Conformal prediction and people (25 comments)<br />
More on the disconnect between who voters support and what they support (25 comments)<br />
Ancestor-worship in academia: Where does it happen? (25 comments)<br />
A welcome rant on betting, knowledge, belief, and the foundations of probability (25 comments)<br />
Oh no Stanford no no no not again please make it stop (25 comments)<br />
Evidence-based Medicine Eats Itself, and How to do Better (my talk at USC this Friday) (25 comments)<br />
This one might possibly be interesting. (25 comments)<br />
4 different meanings of p-value (and how my thinking has changed) (25 comments)<br />
Credit where due to NPR regarding science data fraud, and here’s how they can do even better (25 comments)<br />
Why isn’t Barack Obama out there giving political speeches? (24 comments)<br />
“Exclusive: Embattled dean accused of plagiarism in NSF report” (yup, it’s the torment executioners) (24 comments)<br />
Mindlessness in the interpretation of a study on mindlessness (and why you shouldn’t use the word “whom” in your dating profile) (24 comments)<br />
“Here’s the Unsealed Report Showing How Harvard Concluded That a Dishonesty Expert Committed Misconduct” (24 comments)<br />
Again on the role of elite media in spreading UFOs-as-space-aliens and other bad ideas (24 comments)<br />
The piranha problem: Large effects swimming in a small pond (24 comments)<br />
Crap papers with crude political agendas published in scientific journals: A push-pull problem (24 comments)<br />
log(A + x), not log(1 + x) (24 comments)<br />
B-school prof data sleuth lawsuit fails (24 comments)<br />
Nonsampling error and the anthropic principle in statistics (24 comments)<br />
A 10% swing in win probability corresponds (approximately) to a 0.4% swing in predicted vote (24 comments)<br />
How large is that treatment effect, really? (My talk at the NYU economics seminar, <del>Thurs 7 Mar</del> 18 Apr) (23 comments)<br />
“I was left with an overwhelming feeling that the World Values Survey is simply a vehicle for telling stories about values . . .” (23 comments)<br />
Grappling with uncertainty in forecasting the 2024 U.S. presidential election (23 comments)<br />
Google is violating the First Law of Robotics. (23 comments)<br />
A question for Nate Cohn at the New York Times regarding a claim about adjusting polls using recalled past vote (23 comments)<br />
New Course: Prediction for (Individualized) Decision-making (23 comments)<br />
Bias remaining after adjusting for pre-treatment variables. Also the challenges of learning through experimentation. (23 comments)<br />
“Accounting for Nonresponse in Election Polls: Total Margin of Error” (23 comments)<br />
This post is not really about Aristotle. (22 comments)<br />
The free will to repost (22 comments)<br />
Paper cited by Stanford medical school professor retracted—but even without considering the reasons for retraction, this paper was so bad that it should never have been cited. (22 comments)<br />
“AI” as shorthand for turning off our brains. (This is not an anti-AI post; it’s a discussion of how we think about AI.) (22 comments)<br />
Fewer kids in our future: How historical experience has distorted our sense of demographic norms (22 comments)<br />
More on the oldest famous person ever (just considering those who lived to at least 104) (22 comments)<br />
Freakonomics asks, “Why is there so much fraud in academia,” but without addressing one big incentive for fraud, which is that, if you make grabby enough claims, you can get featured in . . . Freakonomics! (22 comments)<br />
Fake stories in purported nonfiction (22 comments)<br />
Applications of (Bayesian) variational inference? (22 comments)<br />
The immediate victims of the con would rather act as if the con never happened. Instead, they’re mad at the outsiders who showed them that they were being fooled. (21 comments)<br />
The contrapositive of “Politics and the English Language.” One reason writing is hard: (21 comments)<br />
“A passionate group of scientists determined to revolutionize the traditional publishing model in academia” (21 comments)<br />
It’s Ariely time! They had a preregistration but they didn’t follow it. (21 comments)<br />
Do research articles have to be so one-sided? (21 comments)<br />
Adverse Adult Research Outcomes Increased After Increased Willingness of Public Health Journals to Publish Absolute Crap (21 comments)<br />
The NYT sinks to a new low in political coverage (21 comments)<br />
Calibration is sometimes sufficient for trusting predictions. What does this tell us when human experts use model predictions? (21 comments)<br />
Keith O’Rourke’s final published paper: “Statistics as a social activity: Attitudes toward amalgamating evidence” (21 comments)<br />
Close Reading Archive (21 comments)<br />
Ben Shneiderman’s Golden Rules of Interface Design (20 comments)<br />
Clinical trials that are designed to fail (20 comments)<br />
Zotero now features retraction notices (20 comments)<br />
“He had acquired his belief not by honestly earning it in patient investigation, but by stifling his doubts. And although in the end he may have felt so sure about it that he could not think otherwise, yet inasmuch as he had knowingly and willingly worked himself into that frame of mind, he must be held responsible for it.” (20 comments)<br />
Evidence, desire, support (20 comments)<br />
Now here’s a tour de force for ya (20 comments)<br />
To what extent is psychology different from other fields regarding fraud and replication problems? (20 comments)<br />
What to do with election forecasts after Biden is replaced on the ticket? Also something on prediction markets. (20 comments)<br />
The Rise and Fall of the Rock Stars (20 comments)<br />
Carroll/Langer: Credulous, scientist-as-hero reporting from a podcaster who should know better (20 comments)<br />
Who understands alignment anyway (19 comments)<br />
“‘Pure Craft’ Is a Lie” and other essays by Matthew Salesses (19 comments)<br />
Myths of American history from the left, right, and center; also a discussion of the Why everything you thought you knew was wrong” genre of book. (19 comments)<br />
Sorry, NYT, but, yes, “Equidistant Letter Sequences in the Book of Genesis” was junk science (19 comments)<br />
Clarke’s Law, and who’s to blame for bad science reporting (18 comments)<br />
Preregistration is a floor, not a ceiling. (18 comments)<br />
Hey! Here’s a study where all the preregistered analyses yielded null results but it was presented in PNAS as being wholly positive. (18 comments)<br />
“Andrew, you are skeptical of pretty much all causal claims. But wait, causality rules the world around us, right? Plenty have to be true.” (18 comments)<br />
Not eating sweet potatoes: Is that gonna kill me? (18 comments)<br />
Three takes on the protests at Columbia University (18 comments)<br />
Marquand. (18 comments)<br />
5 different reasons why it’s important to include pre-treatment variables when designing and analyzing a randomized experiment (or doing any causal study) (18 comments)<br />
How did the press do on that “black spatula” story? Not so great. (18 comments)<br />
Here’s a sad post for you to start the new year. The Onion (ok, an Onion-affiliate site) is plagiarizing. For reals. (17 comments)<br />
Jonathan Bailey vs. Stephen Wolfram (17 comments)<br />
The data are on a 1-5 scale, the mean is 4.61, and the standard deviation is 1.64 . . . What’s so wrong about that?? (17 comments)<br />
What to make of implicit biases in LLM output? (17 comments)<br />
Statistics Blunder at the Supreme Court (17 comments)<br />
Design analysis is not just about statistical significance and power; it’s relevant for Bayesian inference too. (17 comments)<br />
What can aspiring political moderates learn from the example of Nelson Rockefeller? (17 comments)<br />
Chutzpah is their superpower (Dominic Sandbrook edition) (17 comments)<br />
Stan’s autodiff is 4x faster than JAX on CPU but 5x slower on GPU (in one eval) (17 comments)<br />
“Toward reproducible research: Some technical statistical challenges” and “The political content of unreplicable research” (my talks at Berkeley and Stanford this Wed and Thurs) (17 comments)<br />
Clybourne Park. And a Jamaican beef patty. (But no Gray Davis, no Grover Norquist, no rabbi.) (17 comments)<br />
Columbia Surgery Prof Fake Data Update . . . (yes, he’s still being promoted on the university webpage) (17 comments)<br />
Objects of the class “David Owen” (17 comments)<br />
Why does this guy have 2 gmail accounts? (17 comments)<br />
Why am I willing to bet you $100-1000 there will be a Nobel Prize for Adaptive Experimentation in the next 40 years? (17 comments)<br />
Regarding the use of “common sense” when evaluating research claims (16 comments)<br />
“Whistleblowers always get punished” (16 comments)<br />
Every time Tyler Cowen says, “Median voter theorem still underrated! Hail Anthony Downs!”, I’m gonna point him to this paper . . . (16 comments)<br />
People have needed rituals to turn data into truth for many years. Why would we be surprised if many people now need procedural reforms to work? (16 comments)<br />
No, it’s not “statistically implausible” when results differ between studies, or between different groups within a study. (16 comments)<br />
I just got a strange phone call from two people who claimed to be writing a news story. They were asking me very vague questions and I think it was some sort of scam. I guess this sort of thing is why nobody answers the phone anymore. (16 comments)<br />
“A bizarre failure in the review process at PNAS” (16 comments)<br />
From what body part does the fish rot? (16 comments)<br />
Put multiple graphs on a page: that’s what Nathan Yau says, and I agree. (16 comments)<br />
Remember that paper that reported contagion of obesity? How’s it being cited nowadays? (16 comments)<br />
It’s martingale time, baby! How to evaluate probabilistic forecasts before the event happens? Rajiv Sethi has an idea. (Hint: it involves time series.) (16 comments)<br />
The odd non-spamness of some spam comments (16 comments)<br />
“And while I don’t really want a back-and-forth . . .” (15 comments)<br />
“When will AI be able to do scientific research both cheaper and better than us, thus effectively obsoleting humans?” (15 comments)<br />
Scientific publishers busily thwarting science (again) (15 comments)<br />
A suggestion on how to improve the broader impacts statement requirement for AI/ML papers (15 comments)<br />
Putting a price on vaccine hesitancy (Bayesian analysis of a conjoint experiment) (15 comments)<br />
What is your superpower? (15 comments)<br />
Who is the Stephen Vincent Benet of today? (15 comments)<br />
“The Active Seating Zone (An Educational Experiment)” (15 comments)<br />
The JAMA effect plus news media echo chamber: More misleading publicity on that problematic claim that lesbians and bisexual women die sooner than straight women (15 comments)<br />
Heroes and Villains: The Effects of Identification Strategies on Strong Causal Claims in France (15 comments)<br />
Present Each Other’s Posters: An update after 15 years (15 comments)<br />
Different perspectives on the claims in the paper, The Colonial Origins of Comparative Development (15 comments)<br />
What makes an MCMC sampler GPU-friendly? (15 comments)<br />
Meta-analysis with a single study (15 comments)<br />
Those correction notices, in full. (Yes, it’s possible to directly admit and learn from error.) (15 comments)<br />
Practical issues with calibration for every group and every decision problem (15 comments)<br />
Presidential campaign effects are small. (15 comments)<br />
Bayesian inference (and mathematical reasoning more generally) isn’t just about getting the answer; it’s also about clarifying the mapping from assumptions to inference to decision. (15 comments)<br />
The Lakatos soccer training (14 comments)<br />
“Science as Verified Trust” (14 comments)<br />
Hand-drawn Statistical Workflow at Nelson Mandela (14 comments)<br />
Combining multiply-imputed datasets, never easy (14 comments)<br />
Simulation from a baseline model as a way to better understand your data: This is what “hypothesis testing” should be. (14 comments)<br />
Age gaps between spouses in U.S., U.K., and India (14 comments)<br />
“Alphabetical order of surnames may affect grading” (14 comments)<br />
Piranhas for “omics”? (14 comments)<br />
Pete Rose and gambling addiction: An insight and a question (14 comments)<br />
Update on that politically-loaded paper published in Demography that I characterized as a “hack job”: Further post-publication review (14 comments)<br />
Bad science as genre fiction: I think there’s a lot to be said for this analogy! (14 comments)<br />
If you wanted to be a top tennis player in the late 1930s, there was a huge benefit to being a member of ____. Or to being named ____. (14 comments)<br />
The Village Voice in the 1960s/70s and blogging in the early 2000s (14 comments)<br />
Social penumbras predict political attitudes (my talk at Harvard on Monday Feb 12 at noon) (13 comments)<br />
A new piranha paper (13 comments)<br />
“There is a war between the ones who say there is a war, and the ones who say there isn’t.” (13 comments)<br />
Another opportunity in MLB for Stan users: the Phillies are hiring (13 comments)<br />
Statistical factuality versus practicality versus poetry (13 comments)<br />
Nicholas Carlini on LLMs and AI for research programmers (13 comments)<br />
Movements in the prediction markets, and going beyond a black-box view of markets and prediction models (13 comments)<br />
ChatGPT o1-preview can code Stan (13 comments)<br />
What if the polls are right? (some scatterplots, and some comparisons to vote swings in past decades) (13 comments)<br />
Help teaching short-course that has a healthy dose of data simulation (13 comments)<br />
Answering two questions, one about Bayesian post-selection inference and one about prior and posterior predictive checks (13 comments)<br />
Evaluating samplers with reference draws (12 comments)<br />
Refuted papers continue to be cited more than their failed replications: Can a new search engine be built that will fix this problem? (12 comments)<br />
Their signal-to-noise ratio was low, so they decided to do a specification search, use a one-tailed test, and go with a p-value of 0.1. (12 comments)<br />
You probably don’t have a general algorithm for an MLE of Gaussian mixtures (12 comments)<br />
This is a very disturbing map. (12 comments)<br />
This is not an argument against self-citations. It’s an argument about how they should be counted. Also, a fun formula that expresses the estimated linear regression coefficient as a weighted average of local slopes. (12 comments)<br />
Data issues in that paper that claims that TikTok and Instagram have consumption spillovers that lead to negative utility (12 comments)<br />
Sports media &gt; Prestige media (space aliens edition) (12 comments)<br />
In search of a theory associating honest citation with a higher/deeper level of understanding than (dishonest) plagiarism (12 comments)<br />
Sports gambling addiction epidemic fueled by some combination of psychology, economics, and politics (12 comments)<br />
Awesome online graph guessing game. And scatterplot charades. (12 comments)<br />
“Pitfalls of Demographic Forecasts of US Elections” (12 comments)<br />
A feedback loop can destroy correlation: This idea comes up in many places. (11 comments)<br />
The paradox of replication studies: A good analyst has special data analysis and interpretation skills. But it’s considered a bad or surprising thing that if you give the same data to different analysts, they come to different conclusions. (11 comments)<br />
Hey, here’s some free money for you! Just lend your name to this university and they’ll pay you $1000 for every article you publish! (11 comments)<br />
Mary Rosh! (11 comments)<br />
When Steve Bannon meets the Center for Open Science: Bad science and bad reporting combine to yield another ovulation/voting disaster (11 comments)<br />
Boris and Natasha in America: How often is the wife taller than the husband? (11 comments)<br />
What happens when you’ve had deferential media coverage and then, all of a sudden, you’re treated as a news item rather than as a figure of admiration? (11 comments)<br />
A data science course for high school students (11 comments)<br />
Is there a balance to be struck between simple hierarchical models and more complex hierarchical models that augment the simple frameworks with more modeled interactions when analyzing real data? (11 comments)<br />
Blog was down and is now operating again. (11 comments)<br />
The state of statistics in 1990 (11 comments)<br />
“A Hudson Valley Reckoning: Discovering the Forgotten History of Slaveholding in My Dutch American Family” (11 comments)<br />
Probabilistic numerics and the folk theorem of statistical computing (11 comments)<br />
Specification curve analysis and the multiverse (11 comments)<br />
The true meaning of the alzabo (11 comments)<br />
Resources for teaching and learning survey sampling, from Scott Keeter at Pew Research (10 comments)<br />
Hey! Here’s some R code to make colored maps using circle sizes proportional to county population. (10 comments)<br />
Our new book, Active Statistics, is now available! (10 comments)<br />
“Hot hand”: The controversy that shouldn’t be. And thinking more about what makes something into a controversy: (10 comments)<br />
N=43, “a statistically significant 226% improvement,” . . . what could possibly go wrong?? (10 comments)<br />
If I got a nickel every time . . . (10 comments)<br />
“Nonreplicable” publications are cited more than “replicable” ones? (10 comments)<br />
Implicit assumptions in the Tversky/Kahneman example of the blue and green taxicabs (10 comments)<br />
It’s Stanford time, baby: 8-hour time-restricted press releases linked to a 91% higher risk of hype (10 comments)<br />
“Responsibility for Raw Data”: “Failure to retain data for some reasonable length of time following publication would produce notoriety equal to the notoriety attained by publishing inaccurate results. A possibly more effective means of controlling quality of publication would be to institute a system of quality control whereby random samples of raw data from submitted journal articles would be requested by editors and scrutinized for accuracy and the appropriateness of the analysis performed.” (10 comments)<br />
The interactions paradox in statistics (10 comments)<br />
Here is the Data Sharing Statement, in its entirety, for van Dyck CH, Swanson CJ, Aisen P, et al. Trial of Lecanemab in Early Alzheimer’s Disease. N Engl J Med. DOI: 10.1056/NEJMoa2212948. (10 comments)<br />
Which book should you read first, Active Statistics or Regression and Other Stories? (10 comments)<br />
Stan Playground: Run Stan on the web, play with your program and data at will, and no need to download anything on your computer (10 comments)<br />
How to cheat at Codenames; cheating at board games more generally (10 comments)<br />
Progress in 2023 (9 comments)<br />
What’s up with spring blooming? (9 comments)<br />
Fun with Dååta: Reference librarian edition (9 comments)<br />
Hey, I got tagged by RetractoBot! (9 comments)<br />
Minimum criteria for studies evaluating human decision-making (9 comments)<br />
Here’s some academic advice for you: Never put your name on a paper you haven’t read. (9 comments)<br />
Defining optimal reliance on model predictions in AI-assisted decisions (9 comments)<br />
Philip K. Dick’s character names (9 comments)<br />
Evilicious 3: Face the Music (9 comments)<br />
Two kings, a royal, a knight, and three princesses walk into a bar . . . (Dude from Saudi Arabia accuses the lords of AI of not giving him enough credit.) (9 comments)<br />
Dan Luu asks, “Why do people post on [bad platform] instead of [good platform]?” (9 comments)<br />
When the story becomes the story (9 comments)<br />
1. Why so many non-econ papers by economists? 2. What’s on the math GRE and what does this have to do with stat Ph.D. programs? 3. How does modern research on combinatorics relate to statistics? (9 comments)<br />
Pervasive randomization problems, here with headline experiments (9 comments)<br />
Some solid criticisms of Ariely and Nudge—from 2012! (9 comments)<br />
It’s lumbar time: Wrong inference because of conditioning on a reasonable, but in this case false, assumption. (9 comments)<br />
You can guarantee that the term “statistical guarantee” will irritate me. Here’s why, and let’s go into some details. (9 comments)<br />
It’s $ time! How much should we charge for a link? (9 comments)<br />
“How bad are search results?” Dan Luu has some interesting thoughts: (9 comments)<br />
Some books: The Good Word (1978), The Hitler Conspiracies (2020), In Defense of History (1999), The Book of the Month (1986), Slow Horses (2010), Freedom’s Dominion (2022), A Meaningful Life (1971) (9 comments)<br />
Background on “fail fast” (9 comments)<br />
“Unusual Betting Patterns With Several Temple Games”: It’s martingale time, baby! (9 comments)<br />
“The Stadium” by Frank Guridy (9 comments)<br />
“Very interesting failed attempt at manipulation on Polymarket today” (9 comments)<br />
Supercentenarian superfraud update (9 comments)<br />
Inaccuracy in New York magazine report on election forecasting (9 comments)<br />
The comments section: A request to non-commenters, occasional commenters, and frequent commenters (9 comments)<br />
20-year anniversary of this blog (9 comments)<br />
Supporting Bayesian modeling workflows with iterative filtering for multiverse analysis (9 comments)<br />
That day in 1977 when Jerzy Neyman committed the methodological attribution fallacy. (9 comments)<br />
Plagiarism searches and post-publication review (9 comments)<br />
Physics is like Brazil, Statistics is like Chile (9 comments)<br />
Progress in 2023, Aki’s software edition (8 comments)<br />
Learning from mistakes (my online talk for the American Statistical Association, 2:30pm Tues 30 Jan 2024) (8 comments)<br />
Lefty Driesell and Bobby Knight (8 comments)<br />
There is no golden path to discovery. One of my problems with all the focus on p-hacking, preregistration, harking, etc. is that I fear that it is giving the impression that all will be fine if researchers just avoid “questionable research practices.” And that ain’t the case. (8 comments)<br />
How large is that treatment effect, really? (my talk at NYU economics department Thurs 18 Apr 2024, 12:30pm) (8 comments)<br />
Delayed retraction sampling (8 comments)<br />
“Close but no cigar” unit tests and bias in MCMC (8 comments)<br />
Infovis, infographics, and data visualization: My thoughts 12 years later (8 comments)<br />
6 ways to follow this blog (8 comments)<br />
For that price he could’ve had 54 Jamaican beef patties or 1/216 of a conference featuring Gray Davis, Grover Norquist, and a rabbi (8 comments)<br />
Break it to grok it: The best way to understand how a method works is go construct scenarios where it fails (8 comments)<br />
Loving, hating, and sometimes misinterpreting conformal prediction for medical decisions (8 comments)<br />
Here’s a useful response by Christakis to criticisms of the contagion-of-obesity claims (8 comments)<br />
Here is the Data Sharing Statement, in its entirety, for Goodwin GM, Aaronson ST, Alvarez O, et al. Single-Dose Psilocybin for a Treatment-Resistant Episode of Major Depression. N Engl J Med. DOI: 10.1056/NEJMoa2206443. (8 comments)<br />
Progress in 2023, Jessica Edition (7 comments)<br />
Storytelling and Scientific Understanding (my talks with Thomas Basbøll at Johns Hopkins on 26 Apr) (7 comments)<br />
Listen to those residuals (7 comments)<br />
When do we expect conformal prediction sets to be helpful? (7 comments)<br />
Hey! A new (to me) text message scam! Involving a barfing dog! (7 comments)<br />
Michael Lewis. (7 comments)<br />
Free online book by Bruno Nicenboim, Daniel Schad, and Shravan Vasishth on Bayesian inference and hierarchical modeling using brms and Stan (7 comments)<br />
Minor-league Stats Predict Major-league Performance, Sarah Palin, and Some Differences Between Baseball and Politics (7 comments)<br />
How to code and impute income in studies of opinion polls? (7 comments)<br />
How often is there a political candidate such as Vivek Ramaswamy who is so much stronger in online polls than telephone polls? (7 comments)<br />
Banning the use of common sense in data analysis increases cases of research failure: evidence from Sweden (7 comments)<br />
“Bayesian Workflow: Some Progress and Open Questions” and “Causal Inference as Generalization”: my two upcoming talks at CMU (7 comments)<br />
Decorative statistics and historical records (7 comments)<br />
Some fun basketball graphs (7 comments)<br />
“A Columbia Surgeon’s Study Was Pulled. He Kept Publishing Flawed Data.” . . . and it appears that he’s still at Columbia! (7 comments)<br />
He wants to compute “the effect of a predictor” (that is, an average predictive comparison) for a hierarchical mixture model. You can do it in Stan! (7 comments)<br />
Luck vs. skill in poker (7 comments)<br />
Basu’s Bears (Fat Bear Week and survey calibration) (7 comments)<br />
Flatiron Institute hiring: postdocs, joint faculty, and permanent research positions (7 comments)<br />
Violent science teacher makes ridiculously unsupported research claims, gets treated by legislatures/courts/media as expert on the effects of homeschooling (7 comments)<br />
Should pollsters preregister their design, data collection, and analyses? (7 comments)<br />
Calibration for everyone and every decision problem, maybe (7 comments)<br />
Iterative imputation and incoherent Gibbs sampling (7 comments)<br />
Data manipulation in the world of long-distance swimming! (7 comments)<br />
Announcing two new members of our blogging team . . . (7 comments)<br />
Progress in 2023, Aki Edition (6 comments)<br />
Michael Wiebe has several new replications written up on his site. (6 comments)<br />
The importance of measurement, and how you can draw ridiculous conclusions from your statistical analyses if you don’t think carefully about measurement . . . Leamer (1983) got it. (6 comments)<br />
Cherry blossoms—not just another prediction competition (6 comments)<br />
Tutorial on varying-intercept, varying-slope multilevel models in Stan, from Will Hipson (6 comments)<br />
Mitzi’s and my talks in Trieste 3 and 4 June 2024 (yes, they’ll be broadcast) (6 comments)<br />
One way you can understand people is to look at where they prefer to see complexity. (6 comments)<br />
Edward Kennedy on the Facebook/Instagram 2020 election experiments (6 comments)<br />
Last week’s summer school on probabilistic AI (6 comments)<br />
Toward a Shnerbian theory that establishes connections between the complexity (nonlinearity, chaotic dynamics, number of components) of a system and the capacity to infer causality from datasets (6 comments)<br />
The “fail fast” principle in statistical computing (6 comments)<br />
Two job openings, one in New York on data visualization, one near Paris on Bayesian modeling (6 comments)<br />
An apparent paradox regarding hypothesis tests and rejection regions (6 comments)<br />
What’s a generative model? PyMC and Stan edition (6 comments)<br />
“Announcing the 2023 IPUMS Research Award Winners” (6 comments)<br />
Pete Rose (6 comments)<br />
eLife press release: Deterministic thinking led to a nonsensical statement (6 comments)<br />
“Reduce likelihood of a tick bite by 73.6 times”? Forking paths on the Appalachian Trail. (6 comments)<br />
Delicate language for talking about statistical guarantees (6 comments)<br />
It’s about time (5 comments)<br />
A gathering of the literary critics: Louis Menand and Thomas Mallon, meet Jeet Heer (5 comments)<br />
Why we say that honesty and transparency are not enough: (5 comments)<br />
Statistical practice as scientific exploration (5 comments)<br />
Analogy between (a) model checking in Bayesian statistics, and (b) the self-correcting nature of science. (5 comments)<br />
Population forecasting for small areas: an example of learning through a social network (5 comments)<br />
Data challenges with the Local News Initiative mapping project (5 comments)<br />
Lucy is not a nickname. (5 comments)<br />
“The Secret Life of John Le Carré” (5 comments)<br />
Evil scamming fake publishers (5 comments)<br />
The Mets are looking to hire a data scientist (5 comments)<br />
Why art is more forgiving than game design (5 comments)<br />
Salesses: “some writing exercises meant to help students with various elements of craft” (5 comments)<br />
StanCon 2024 Oxford: recorded talks are now released! (5 comments)<br />
Code it! (patterns in data edition) (5 comments)<br />
Softmax is on the log, not the logit scale (5 comments)<br />
“My view is that if I can show that a result was cooked and that doing it correctly does not yield the answer the authors claimed, then the result is discredited. . . . What I hear, instead, is the following . . .” (4 comments)<br />
Lancet-bashing! (4 comments)<br />
Bayesian Analysis with Python (4 comments)<br />
Bayesian inference with informative priors is not inherently “subjective” (4 comments)<br />
Here’s something you should do when beginning a project, and in the middle of a project, and in the end of the project: Clearly specify your goals, and also specify what’s not in your goal set. (4 comments)<br />
“When are Bayesian model probabilities overconfident?” . . . and we’re still trying to get to meta-Bayes (4 comments)<br />
“Often enough, scientists are left with the unenviable task of conducting an orchestra with out-of-tune instruments” (4 comments)<br />
Studying causal inference in the presence of feedback: (4 comments)<br />
They’re trying to get a hold on the jungle of cluster analysis. (4 comments)<br />
“Beyond the black box: Toward a new paradigm of statistics in science” (talks this Thursday in London by Jessica Hullman, Hadley Wickham, and me) (4 comments)<br />
Interactive and Automated Data Analysis: thoughts from Di Cook, Hadley Wickham, Jessica Hullman, and others (4 comments)<br />
(This one’s important:) Looking Beyond the Obvious: Essentialism and abstraction as central to our reasoning and beliefs (4 comments)<br />
“The Waltz of Reason” and a paradox of book reviewing (4 comments)<br />
StanCon 2024… is a wrap! (4 comments)<br />
22 Revision Prompts from Matthew Salesses (4 comments)<br />
Two spans of the bridge of inference (4 comments)<br />
Average predictive comparisons (4 comments)<br />
Gayface Data Replicability Problems (4 comments)<br />
Addressing legitimate counterarguments in a scientific review: The challenge of being an insider (4 comments)<br />
Most popular posts of 2024 (4 comments)<br />
What is the minimum bloggable contribution? (4 comments)<br />
What to trust in the newspaper? Example of “The Simple Nudge That Raised Median Donations by 80%” (3 comments)<br />
Bayesian BATS to advance Bayesian Thinking in STEM (3 comments)<br />
Intro to BridgeStan: The new in-memory interface for Stan (3 comments)<br />
Storytelling and Scientific Understanding (my talks with Thomas Basbøll at Johns Hopkins this Friday) (3 comments)<br />
Evaluating MCMC samplers (3 comments)<br />
Applied modelling in drug development? brms! (3 comments)<br />
GPT today: Buffon’s Needle in Python with plotting (and some jokes) (3 comments)<br />
Comedy and child abuse in literature (3 comments)<br />
Cross validation and pointwise or joint measures of prediction accuracy (3 comments)<br />
A guide to detecting AI-generated images, informed by experiments on people’s ability to detect them (3 comments)<br />
Free Textbook on Applied Regression and Causal Inference (3 comments)<br />
Free Book of Stories, Activities, Computer Demonstrations, and Problems in Applied Regression and Causal Inference (3 comments)<br />
Close reading in literary criticism and statistical analysis (3 comments)<br />
GIST: Now with local step size adaptation for NUTS (3 comments)<br />
Should you always include a varying slope for the lower-level variable involved in a cross-level interaction? (3 comments)<br />
3 levels of fraud: One-time, Linear, and Exponential (3 comments)<br />
Postdoc position at Northwestern on evaluating AI/ML decision support (3 comments)<br />
The marginalization or Jeffreys-Lindley paradox: it’s already been resolved. (3 comments)<br />
They solved the human-statistical reasoning interface back in the 80s (2 comments)<br />
“Replicability &amp; Generalisability”: Applying a discount factor to cost-effectiveness estimates. (2 comments)<br />
Leap Day Special! (2 comments)<br />
Postdoc Opportunity at the HEDCO Institute for Evidence-Based Educational Practice in the College of Education at the University of Oregon (2 comments)<br />
GIST: Gibbs self-tuning for HMC (2 comments)<br />
“Former dean of Temple University convicted of fraud for using fake data to boost its national ranking” (2 comments)<br />
Update on “the hat”: It’s “the spectre,” a single shape that can tile the plane aperiodically but not periodically, and doesn’t require flipping (2 comments)<br />
New online Stan course: 80 videos + hosted live coding environment (2 comments)<br />
In Stan, “~” should be called a “distribution statement,” not a “sampling statement.” (2 comments)<br />
Forking paths and workflow in statistical practice and communication (2 comments)<br />
Doctoral student positions in Bayesian workflow at Aalto, Finland (2 comments)<br />
Election prediction markets: What happens next? (2 comments)<br />
Oregon State Stats Dept. is Hiring (2 comments)<br />
Hey, journalist readers! Does anyone have a contact at NPR? (2 comments)<br />
“My quick answer is that I don’t care much about permutation tests because they are testing a null null hypothesis that is of typically no interest” (1 comments)<br />
“Theoretical statistics is the theory of applied statistics”: A scheduled conference on the topic (1 comments)<br />
Progress in 2023, Charles edition (1 comments)<br />
A question about Lindley’s supra Bayesian method for expert probability assessment (1 comments)<br />
Those annoying people-are-stupid narratives in journalism (1 comments)<br />
ISBA 2024 Satellite Meeting: Lugano, 25–28 June (1 comments)<br />
My NYU econ talk will be Thurs 18 Apr 12:30pm (NOT Thurs 7 Mar) (1 comments)<br />
Is the 2024 New York presidential primary really an “important election”? (1 comments)<br />
Fully funded doctoral student positions in Finland (1 comments)<br />
“Randomization in such studies is arguably a negative, in practice, in that it gives apparently ironclad causal identification (not really, given the ultimate goal of generalization), which just gives researchers and outsiders a greater level of overconfidence in the claims.” (1 comments)<br />
Supporting Bayesian modelling workflows with iterative filtering for multiverse analysis (1 comments)<br />
No, I don’t believe the claim that “Mothers negatively affected by having three daughters and no sons, study shows.” (1 comments)<br />
He has some questions about a career in sports analytics. (1 comments)<br />
Questions and Answers for Applied Statistics and Multilevel Modeling (1 comments)<br />
StanCon 2024: scholarships, sponsors, and other news (1 comments)<br />
19 ways of looking at data science at the singularity, from David Donoho and 17 others (1 comments)<br />
A message to Christian Hesse, mathematician and author of chess books (1 comments)<br />
Online seminar for Monte Carlo Methods++ (1 comments)<br />
Faculty positions at the University of Oregon’s new Data Science department (1 comments)<br />
“Tough choices in election forecasting: All the things that can go wrong” (my webinar this Friday 11am with the Washington Statistical Society) (1 comments)<br />
Call for StanCon 2025+ (1 comments)<br />
What should Yuling include in his course on statistical computing? (1 comments)<br />
Calibration “resolves” epistemic uncertainty by giving predictions that are indistinguishable from the true probabilities. Why is this still unsatisfying? (1 comments)<br />
Since Jeffrey Epstein is in the news again . . . (0 comments)<br />
Postdoc at Washington State University on law-enforcement statistics (0 comments)<br />
Here’s how to subscribe to our new weekly newsletter: (0 comments)<br />
Progress in 2023, Leo edition (0 comments)<br />
Click here to help this researcher gather different takes on making data visualizations for blind people (0 comments)<br />
Using the term “visualization” for non-visual representation of data (0 comments)<br />
Varying slopes and intercepts in Stan: still painful in 2024 (0 comments)<br />
BD corner: I came across this interesting interview with Daniel Clowes on the sources for Monica (0 comments)<br />
Hey, some good news for a change! (Child psychology and Bayes) (0 comments)<br />
A nested helix plot that simultaneously shows events on the scales of centuries, millennia, . . . all the way back to billions of years (0 comments)<br />
Job Ad: Spatial Statistics Group Lead at Oak Ridge National Laboratory (0 comments)<br />
Bayesian Workflow, Causal Generalization, Modeling of Sampling Weights, and Time: My talks at Northwestern University this Friday and the University of Chicago on Monday (0 comments)<br />
Papers on human decision-making under uncertainty in ML venues! We have advice. (0 comments)<br />
New stat podcast just dropped (0 comments)<br />
Faculty and postdoc jobs in computational stats at Newcastle University (UK) (0 comments)<br />
Subscribe to this free newsletter and get a heads-up on our scheduled posts a week early! (0 comments)<br />
“What do we need from a probabilistic programming language to support Bayesian workflow?” (0 comments)<br />
StanCon 2024 is in 32 days! (0 comments)<br />
NeurIPS 2024 workshop on Statistical Frontiers in LLMs and Foundation Models (0 comments)<br />
Faculty positions at the University of California on AI, Inequality, and Society (0 comments)<br />
Modeling Weights to Generalize (my talk this Wed noon at the Columbia University statistics department) (0 comments)<br />
Bayesian social science conference in Amsterdam! Next month! (0 comments)<br />
Postdoc opportunity! to work with me here at Columbia! on Bayesian workflow! for contamination models! With some wonderful collaborators!! (0 comments)<br />
NYT catches up to Statistical Modeling, Causal Inference, and Social Science (0 comments)<br />
Leave-one-out cross validation (LOO) for an astronomy problem (0 comments)<br />
Self-reference and self-reproduction of evidence (0 comments)<br />
The Red Sox are hiring (0 comments)<br />
Faculty positions at Princeton in interdisciplinary data science (0 comments)</p>
<p>Thank you all for your contributions!</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2025/01/01/published-or-accepted-for-publication-in-2024/feed/</wfw:commentRss>
			<slash:comments>7</slash:comments>
		
		
			</item>
		<item>
		<title>Calibration “resolves” epistemic uncertainty by giving predictions that are indistinguishable from the true probabilities. Why is this still unsatisfying?</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/31/calibration-resolves-epistemic-uncertainty-by-giving-predictions-that-are-indistinguishable-from-the-true-probabilities-why-is-this-still-unsatisfying/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/31/calibration-resolves-epistemic-uncertainty-by-giving-predictions-that-are-indistinguishable-from-the-true-probabilities-why-is-this-still-unsatisfying/#comments</comments>
		
		<dc:creator><![CDATA[Jessica Hullman]]></dc:creator>
		<pubDate>Tue, 31 Dec 2024 17:27:07 +0000</pubDate>
				<category><![CDATA[Bayesian Statistics]]></category>
		<category><![CDATA[Decision Analysis]]></category>
		<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Miscellaneous Statistics]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51522</guid>

					<description><![CDATA[This is Jessica. The last day of the year is like a good time for finishing things up, so I figured it’s time for one last post wrapping up some thoughts on calibration.  As my previous posts got into, calibrated prediction uncertainty &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/31/calibration-resolves-epistemic-uncertainty-by-giving-predictions-that-are-indistinguishable-from-the-true-probabilities-why-is-this-still-unsatisfying/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><span style="font-weight: 400">This is Jessica. The last day of the year is like a good time for finishing things up, so I figured it’s time for one last post wrapping up some thoughts on calibration. </span></p>
<p><span style="font-weight: 400">As my </span><a href="https://statmodeling.stat.columbia.edu/2024/08/14/when-is-calibration-enough/"><span style="font-weight: 400">previous</span></a> <a href="https://statmodeling.stat.columbia.edu/2024/11/01/calibration-is-sometimes-sufficient-for-trusting-predictions-what-does-this-tell-us-when-human-experts-use-model-predictions/"><span style="font-weight: 400">posts</span></a> <a href="https://statmodeling.stat.columbia.edu/2024/11/15/calibration-for-everyone-and-every-decision-maybe/"><span style="font-weight: 400">got into</span></a><span style="font-weight: 400">, calibrated prediction uncertainty is the goal of various posthoc calibration algorithms discussed in machine learning research, which use held out data to learn transformations on model predicted probabilities in order to achieve calibration on the held out data. I’ve reflected a bit on what calibration can and can’t give us in terms of assurances for decision-making. Namely, it makes predictions trustworthy for decisions in the restricted sense that a decision-maker who will choose their action purely based on the prediction can’t do better than treating the calibrated predictions as the true probabilities. </span></p>
<p><span style="font-weight: 400">But something I’ve had trouble articulating as clearly as I’d like involves what’s missing (and why) when it comes to what calibration gives us versus a more complete representation of the limits of our knowledge in making some predictions. </span></p>
<p><span style="font-weight: 400">The distinction involves how we express higher order uncertainty. Let’s say we are doing multiclass classification, and fit a model fhat to some labeled data. Our “level 0” prediction fhat(x) contains no uncertainty representation at all; we check it against the ground truth y. Our “level 1” prediction phat(.|x) predicts the conditional distribution over classes; we check it against the empirical distribution that gives a probability p(y|x) for each possible y. Our “level 2” prediction is trying to predict the distribution of the conditional distribution over classes, p(p(.|x), e.g. a Dirichlet distribution that assigns probability to each distribution p(.|x), which we can distinguish using some parameters theta.</span></p>
<p><span style="font-weight: 400">From a Bayesian modeling perspective, it’s natural to think about distributions of distributions. A prior distribution over model parameters implies a distribution over possible data-generating distributions. Upon fitting a model, the posterior predictive distribution summarizes both “aleatoric” uncertainty due to inherent randomness in the generating process and “epistemic” uncertainty stemming from our lack of knowledge of the true parameter values. </span></p>
<p><span style="font-weight: 400">In some sense calibration “resolves” epistemic uncertainty by providing point predictions that are indistinguishable from the true probabilities. But if you’re hoping to get a faithful summary of the current state of knowledge, it can seem like something is still missing. In the Bayesian framework, we can collapse our posterior prediction of the outcome y for any particular input x to a point estimate, but we don’t have to. </span></p>
<p><span style="font-weight: 400">Part of the difficulty is that whenever we evaluate performance as loss over some data-generating distribution, having more than a point estimate is not necessary. This is true even without considering second order uncertainty. If we train a level 0 prediction of the outcome y using the standard loss minimization framework with 0/1 loss, then it will learn to predict the mode. And so to the extent that it’s hard to argue one’s way out of loss minimization as a standard for evaluating decisions, it’s hard to motivate faithful expression of epistemic uncertainty.</span></p>
<p><span style="font-weight: 400">For second order uncertainty, the added complication is there is no ground truth. We might believe there is some intrinsic value in being able to model uncertainty about the best predictor, but how do we formalize this given that there’s no ground truth against which to check our second order predictions? We can’t learn by drawing samples from the distribution that assigns probability to different first order distributions p(.|x) because technically there is no such distribution beyond our conception of it. </span></p>
<p><a href="https://statmodeling.stat.columbia.edu/2024/11/01/calibration-is-sometimes-sufficient-for-trusting-predictions-what-does-this-tell-us-when-human-experts-use-model-predictions/#comment-2382770"><span style="font-weight: 400">Daniel Lakeland previously provided an example</span></a><span style="font-weight: 400"> I found helpful of the difference between putting Bayesian probability on a predicted frequency, where there’s no sense in which we can check the calibration of the second order prediction. </span></p>
<p><span style="font-weight: 400">Related to this, I recently came across a </span><a href="https://arxiv.org/abs/2203.06102"><span style="font-weight: 400">few papers</span></a> <a href="https://arxiv.org/abs/2301.12736"><span style="font-weight: 400">by Viktor Bengs et al</span></a><span style="font-weight: 400"> that formalize some of this in an ML context. Essentially, they show that there is no well-defined loss function that can be used in the typical ML learning pipeline to incentivize the learner to make correct predictions that are also faithful as expressions of the epistemic uncertainty. This can be expressed in terms of trying to find a </span><a href="https://arxiv.org/abs/2301.12736"><span style="font-weight: 400">proper scoring rule</span></a><span style="font-weight: 400">. In the case of first order predictions, as long as we use a proper scoring rule as the loss function, we can expect accurate predictions, because a proper scoring rule is one for which one cannot score higher by deviating from reporting our true beliefs. But there is no loss function that incentivizes a second-order learner to faithfully represent its epistemic uncertainty like a proper scoring rule does for a first order learner. </span></p>
<p><span style="font-weight: 400">This may seem obvious, especially if you’re coming from a Bayesian tradition, considering that there is no ground truth against which to score second order predictions. And yet, various loss functions have been proposed for estimating level 2 predictors in the ML literature, such as minimizing the empirical loss of the level 1 prediction averaged over possible parameter values. These results make clear that one needs to be careful interpreting the predictors they give, because, e.g., they can actually incentivize predictors that appear to be certain about the first order distribution. </span></p>
<p><span style="font-weight: 400">I guess a question that remains is how to talk about incentives for second order uncertainty at all in a context where minimizing loss from predictions is the primary goal. I don’t think the right conclusion is that it doesn’t matter since we can’t integrate it into a loss minimization framework. Having the ability to decompose predictions by different sources of uncertainty and be explicit about what our higher order uncertainty looks like going in (i.e., by defining a prior) has scientific value in less direct ways, like communicating beliefs and debugging when things go wrong. </span></p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/31/calibration-resolves-epistemic-uncertainty-by-giving-predictions-that-are-indistinguishable-from-the-true-probabilities-why-is-this-still-unsatisfying/feed/</wfw:commentRss>
			<slash:comments>5</slash:comments>
		
		
			</item>
		<item>
		<title>What is the minimum bloggable contribution?</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/31/what-is-the-minimum-bloggable-contribution/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/31/what-is-the-minimum-bloggable-contribution/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Tue, 31 Dec 2024 14:41:28 +0000</pubDate>
				<category><![CDATA[Decision Analysis]]></category>
		<category><![CDATA[Literature]]></category>
		<category><![CDATA[Sociology]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=50987</guid>

					<description><![CDATA[The other day someone sent me an email pointing me to an online article, a statistical analysis criticizing an online article that was a reanalysis of data from an article that was a meta-analysis and literature review of a controversial &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/31/what-is-the-minimum-bloggable-contribution/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>The other day someone sent me an email pointing me to an online article, a statistical analysis criticizing an online article that was a reanalysis of data from an article that was a meta-analysis and literature review of a controversial topic that had been written about in some earlier published papers that were themselves literature reviews.</p>
<p>My correspondent was interested in my take, and I replied that the latest article made some good points and also some errors.  I didn&#8217;t think I had anything useful to say on this one so I didn&#8217;t post on the topic.</p>
<p>Given that I&#8217;d already gone to the trouble of reading&#8211;OK, skimming&#8211;all these articles, and I can post here for free, arguably I could contribute in a useful way just with a short post explaining where I agreed with this new article and where I thought it went wrong.  I have no real or perceived beefs with any of the people involved in this one, and I expect that whatever feedback I were to provide would be taken constructively, not <a href="https://statmodeling.stat.columbia.edu/2019/01/18/ladder-responses-criticism-responsible-destructive/">defensively</a>.</p>
<p>So why not post?  The difficulty here would come not directly in my comments on those articles but rather in the necessary scaffolding:  all the bits I&#8217;d need to add to avoid writing something that could be misinterpreted.</p>
<p>One of the challenges of writing, as compared to speaking, is that your words are just out there, interpreted without the benefit of intonation, context, and dialogue.  This isn&#8217;t anything specific to blogging; the same issue can arise when communicating with people by email.  Not that direct face-to-face conversation works either; it&#8217;s just that something written is just out there in some way.</p>
<p>So it&#8217;s a cost-benefit calculation:  weighing my contributions to this particular discussion against the effort of constructing the scaffolding.  In this case I decided the best option was to not bother.</p>
<p>But then I became interested in the meta-topic of when to post, so I wrote this, which I&#8217;ll schedule to appear in a few months.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/31/what-is-the-minimum-bloggable-contribution/feed/</wfw:commentRss>
			<slash:comments>4</slash:comments>
		
		
			</item>
		<item>
		<title>Most popular posts of 2024</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/31/most-popular-posts-of-2024/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/31/most-popular-posts-of-2024/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Tue, 31 Dec 2024 14:12:58 +0000</pubDate>
				<category><![CDATA[Sociology]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51518</guid>

					<description><![CDATA[John Cook writes that he looked on Hacker News to see which of his posts were most popular: &#8220;I [Cook] didn’t look at my server logs, but generally the posts that get the most traffic are posts that someone submits &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/31/most-popular-posts-of-2024/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>John Cook <a href="https://www.johndcook.com/blog/2024/12/24/most-popular-2024/">writes that</a> he looked on Hacker News to see which of his posts were most popular:  &#8220;I [Cook] didn’t look at my server logs, but generally the posts that get the most traffic are posts that someone submits to Hacker News.&#8221;</p>
<p>I don&#8217;t look at my server logs either&#8211;that way lies madness!  My question is, how much should we trust the popularity ranking from Hacker News?</p>
<p>I don&#8217;t know much about Hacker News myself.  I <a href="https://news.ycombinator.com">went there</a> and entered statmodeling in the search box, and by default it lists in order of popularity.  Here are the first two pages of the most popular from 2024:</p>
<blockquote><p>Suspicious data pattern in recent Venezuelan election (https://statmodeling.stat.columbia.edu/2024/07/31/suspicious-data-pattern-in-recent-venezuelan-election/)<br />
903 points|kgwgk|5 months ago|470 comments</p>
<p>Bayesian Statistics: The three cultures (https://statmodeling.stat.columbia.edu/2024/07/10/three-cultures-bayes-subjective-objective-pragmatic/)<br />
309 points|luu|5 months ago|109 comments</p>
<p>The immediate victims of a con would rather act as if the con never happened (https://statmodeling.stat.columbia.edu/2024/01/07/french-bio-lab-research-scandal/)<br />
230 points|Tomte|1 year ago|140 comments</p>
<p>Bad stuff going down at the American Sociological Association (https://statmodeling.stat.columbia.edu/2024/01/17/bad-stuff-going-down-at-the-american-sociological-association/)<br />
129 points|Tomte|1 year ago|116 comments</p>
<p>Those correction notices, in full (https://statmodeling.stat.columbia.edu/2024/11/24/those-correction-notices-in-full-yes-its-possible-to-directly-admit-and-learn-from-error/)<br />
128 points|Tomte|1 month ago|35 comments</p>
<p>Dean of Engineering at University of Nevada wrote a paper that’s bad (https://statmodeling.stat.columbia.edu/2024/02/06/its-bezzle-time-the-dean-of-engineering-at-the-university-of-nevada-gets-paid-372127-a-year-and-wrote-a-paper-thats-so-bad-you-cant-believe-it-i-mean-really-you-have-to-take-a-look-at-t/)<br />
123 points|Tomte|11 months ago|79 comments</p>
<p>Defining Statistical Models in Jax? (https://statmodeling.stat.columbia.edu/2024/10/08/defining-statistical-models-in-jax/#comments)<br />
118 points|hackandthink|3 months ago|18 comments</p>
<p>Critique of Freakonomics interview with psychologist Ellen Langer (https://statmodeling.stat.columbia.edu/2024/10/28/freakonomics-does-it-again-not-in-a-good-way-jeez-these-guys-are-credulous/)<br />
94 points|nabla9|2 months ago|82 comments</p>
<p>Refuted papers continue to be cited more than their failed replications (https://statmodeling.stat.columbia.edu/2024/03/10/refuted-papers-continue-to-be-cited-more-than-their-failed-replications-can-a-new-search-engine-be-built-that-will-fix-this-problem/)<br />
79 points|nabla9|10 months ago|34 comments</p>
<p>Well-known paradox of R-squared is still buggin me (https://statmodeling.stat.columbia.edu/2024/06/17/this-well-known-paradox-of-r-squared-is-still-buggin-me-can-you-help-me-out/)<br />
71 points|luu|6 months ago|103 comments</p>
<p>Levels of fraud: One-time, Linear, and Exponential (https://statmodeling.stat.columbia.edu/2024/09/27/3-levels-of-fraud-one-time-linear-and-exponential/)<br />
4 points|Tomte|3 months ago|0 comments</p>
<p>You can guarantee that the term &#8220;statistical guarantee&#8221; will irritate me (https://statmodeling.stat.columbia.edu/2024/07/05/you-can-guarantee-that-the-term-statistical-guarantee-will-irritate-me-heres-why-and-lets-go-into-some-details/)<br />
4 points|Tomte|6 months ago|0 comments</p>
<p>Is it really &#8220;the economy, stupid&#8221;? (https://statmodeling.stat.columbia.edu/2024/05/11/is-it-really-the-economy-stupid/)<br />
4 points|munichpavel|8 months ago|0 comments</p>
<p>What are the best scientific papers ever written? (https://statmodeling.stat.columbia.edu/2020/06/09/what-are-the-best-scientific-papers-ever-written/)<br />
4 points|reqo|9 months ago|0 comments</p>
<p>Zotero now features retraction notices (https://statmodeling.stat.columbia.edu/2024/03/12/zotero-now-features-retraction-notices/)<br />
4 points|Tomte|10 months ago|0 comments</p>
<p>&#8220;the income of the average American will double approximately every 39 years&#8221;? (https://statmodeling.stat.columbia.edu/2024/05/15/the-income-of-the-average-american-will-double-approximately-every-39-years-who-says-that-sort-of-thing-show-some-respect-for-uncertainty-dude/)<br />
3 points|luu|7 months ago|2 comments</p>
<p>Prediction markets and the need for &#8220;dumb money&#8221; as well as &#8220;smart money&#8221; (https://statmodeling.stat.columbia.edu/2024/10/25/prediction-markets-and-the-need-for-dumb-money-as-well-as-smart-money/)<br />
3 points|nabla9|2 months ago|1 comments</p>
<p>If you want to play tennis at the top level there&#8217;s a benefit to being ____ (https://statmodeling.stat.columbia.edu/2024/09/06/meritocracy-and-womens-tennis/)<br />
3 points|noelwelsh|4 months ago|1 comments</p>
<p>Why bother engaging outside reviewers at all? (https://statmodeling.stat.columbia.edu/2024/12/14/the-most-interesting-part-of-the-story-is-that-the-publisher-went-through-all-these-steps-of-reviewing-and-revising-if-they-just-want-to-make-money-by-publishing-crap-why-bother-engaging-outside-re/)<br />
3 points|Tomte|16 days ago|0 comments</p>
<p>ChatGPT o1-preview can code Stan (https://statmodeling.stat.columbia.edu/2024/10/22/chatgpt-o1-preview-can-code-stan/)<br />
3 points|lr0|2 months ago|0 comments</p>
<p>Sean Carroll/Ellen Langer: Credulous, scientist-as-hero reporting (https://statmodeling.stat.columbia.edu/2024/10/19/carroll-langer-credulous-scientist-as-hero-reporting-from-a-podcaster-who-should-know-better/)<br />
3 points|nabla9|2 months ago|0 comments</p>
<p>Freakonomics asks, &#8220;Why is there so much fraud in academia,&#8221; missing one reason (https://statmodeling.stat.columbia.edu/2024/09/14/freakonomics-asks-why-is-there-so-much-fraud-in-academia-but-without-addressing-one-big-incentive-for-fraud-which-is-that-if-you-make-grabby-enough-claims-you-can-get-features-in-freako/)<br />
3 points|nabla9|4 months ago|0 comments</p>
<p>Google is violating the First Law of Robotics (https://statmodeling.stat.columbia.edu/2024/07/16/google-is-violating-the-first-law-of-robotics/)<br />
3 points|nabla9|6 months ago|0 comments</p>
<p>Best way to understand how a method works? Construct scenarios where it fails (https://statmodeling.stat.columbia.edu/2024/05/25/break-it-to-grok-it-the-best-way-to-understand-how-a-method-works-is-go-construct-scenarios-where-it-fails/)<br />
3 points|rossdavidh|7 months ago|0 comments</p>
<p>Infovis, Infographics, and Data Visualization (https://statmodeling.stat.columbia.edu/2024/04/18/infovis-infographics-and-data-visualization-my-thoughts-12-years-later/)<br />
3 points|Tomte|9 months ago|0 comments</p>
<p>What is the prevalence of bad social science? (https://statmodeling.stat.columbia.edu/2024/04/06/what-is-the-prevalence-of-bad-social-science/)<br />
3 points|Tomte|9 months ago|0 comments</p>
<p>The Contrapositive of &#8220;Politics and the English Language.&#8221; (https://statmodeling.stat.columbia.edu/2024/03/25/the-contrapositive-of-politics-and-the-english-language-one-reason-writing-is-hard/)<br />
3 points|luu|9 months ago|0 comments</p>
<p>The Contrapositive of &#8220;Politics and the English Language.&#8221; (https://statmodeling.stat.columbia.edu/2024/03/25/the-contrapositive-of-politics-and-the-english-language-one-reason-writing-is-hard/)<br />
3 points|Tomte|9 months ago|0 comments</p>
<p>How can a top scientist be so confidently wrong? (2022) (https://statmodeling.stat.columbia.edu/2022/06/08/how-can-a-top-scientist-be-so-confidently-wrong-r-a-fisher-and-smoking-example/)<br />
3 points|EndXA|10 months ago|0 comments</p>
<p>Michael Lewis (https://statmodeling.stat.columbia.edu/2024/02/22/michael-lewis/)<br />
3 points|luu|10 months ago|0 comments</p>
<p>Is Matthew Walker&#8217;s &#8220;Why We Sleep&#8221; Riddled with Scientific/Factual Errors? (2019) (https://statmodeling.stat.columbia.edu/2019/11/18/is-matthew-walkers-why-we-sleep-riddled-with-scientific-and-factual-errors/)<br />
3 points|Tomte|10 months ago|0 comments</p>
<p>Clinical trials that are designed to fail (https://statmodeling.stat.columbia.edu/2024/02/11/clinical-trials-that-are-designed-to-fail/)<br />
3 points|Tomte|11 months ago|0 comments</p>
<p>Number of inline code comments is zero. Nada. Zilch. Nil. Naught (https://statmodeling.stat.columbia.edu/2024/02/07/when-all-else-fails-add-a-code-comment/)<br />
3 points|thaumasiotes|11 months ago|0 comments</p>
<p>The Onion (ok, an Onion-affiliate site) is plagiarizing. For reals (https://statmodeling.stat.columbia.edu/2024/01/01/the-onion-ok-an-onion-affiliate-site-is-plagiarizing/)<br />
3 points|Tomte|1 year ago|0 comments</p>
<p>The rise and fall of Seth Roberts and the Shangri-La diet (2023) (https://statmodeling.stat.columbia.edu/2023/11/20/the-rise-and-fall-of-seth-roberts-and-the-shangri-la-diet/)<br />
2 points|sowbug|1 month ago|1 comments</p>
<p>Prediction markets need dumb money as well as smart money (https://statmodeling.stat.columbia.edu/2024/10/25/prediction-markets-and-the-need-for-dumb-money-as-well-as-smart-money/)<br />
2 points|johndcook|2 months ago|1 comments</p>
<p>Whassup with those economists who predicted a recession that then didn&#8217;t happen? (https://statmodeling.stat.columbia.edu/2024/05/31/whassup-with-those-economists-who-predicted-a-recession-that-then-didnt-happen/)<br />
2 points|_pfco|7 months ago|1 comments</p>
<p>Harvard time baby: completely botched data but you don&#8217;t change your conclusions (https://statmodeling.stat.columbia.edu/2024/12/12/harvard-lexicon-kerfuffle-is-what-you-call-it-when-you-completely-botched-your-data-but-you-dont-want-to-change-your-conclusions/)<br />
2 points|nabla9|17 days ago|0 comments</p>
<p>What it takes to conclude that a research seam has been mined to exhaustion (https://statmodeling.stat.columbia.edu/2024/11/26/i-wonder-just-what-it-takes-to-get-people-to-conclude-that-a-research-seam-has-been-mined-to-the-point-of-exhaustion/)<br />
2 points|domofutu|1 month ago|0 comments</p>
<p>Andrew Gelman is not the science police because there is no science police (https://statmodeling.stat.columbia.edu/2024/11/21/andrew-gelman-is-not-the-science-police-because-there-is-no-such-thing-as-the-science-police/)<br />
2 points|Tomte|1 month ago|0 comments</p></blockquote>
<p>I guess I should be thanking kgwgk, luu, nabla9, Tomte, and the rest of the gang for linking to us!  It&#8217;s interesting how many of the links come from the same few Hacker News participants.</p>
<p>Also, I don&#8217;t know what these &#8220;points&#8221; represent.  2 points doesn&#8217;t sound like a lot, which is kind of a bummer . . . I put so much effort into writing these posts, I&#8217;d always appreciate more readers!</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/31/most-popular-posts-of-2024/feed/</wfw:commentRss>
			<slash:comments>5</slash:comments>
		
		
			</item>
		<item>
		<title>A very interesting discussion by Roy Sorensen of the interesting-number paradox</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/30/a-very-interesting-discussion-by-roy-sorenson-of-the-interesting-number-paradox/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/30/a-very-interesting-discussion-by-roy-sorenson-of-the-interesting-number-paradox/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Mon, 30 Dec 2024 14:09:00 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51517</guid>

					<description><![CDATA[There&#8217;s this mathematical joke that all numbers&#8211;more precisely, all positive integers&#8211;are interesting. As Roy Sorensen puts it: Mathematicians are fond of Edwin Beckenbach’s (1945) argument: A. If some [positive] integer is not interesting, then there is a least such integer. &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/30/a-very-interesting-discussion-by-roy-sorenson-of-the-interesting-number-paradox/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>There&#8217;s this mathematical joke that all numbers&#8211;more precisely, all positive integers&#8211;are interesting. As Roy Sorensen <a href="http://fitelson.org/few/few_08/sorensen.pdf">puts it</a>:</p>
<blockquote><p>Mathematicians are fond of Edwin Beckenbach’s (1945) argument:</p>
<p>A. If some [positive] integer is not interesting, then there is a least such integer.</p>
<p>B. If some integer is the first uninteresting integer, then that fact makes the integer interesting.</p>
<p>C. Therefore, all integers are interesting.</p>
<p>The sophism has attracted no philosophical commentary because of a trivializing resemblance to Berry’s paradox and the sorites.</p></blockquote>
<p>I&#8217;d not heard of either Berry’s paradox or the sorites&#8211;you can look them up yourself on wikipedia. Indeed, you can find <a href="https://en.wikipedia.org/wiki/List_of_paradoxes#Self–reference">a whole list</a> of paradoxes of self-reference.</p>
<p>Sorensen&#8217;s article is fun and thought-provoking. The full reference is Roy Sorensen (2011). Interestingly dull numbers. Philosophy and Phenomenological Research 82, 655-673.</p>
<p>Here&#8217;s how he resolves the paradox:</p>
<blockquote><p>My main objection is to premise B (`If some integer is the first uninteresting integer, then that fact makes the integer interesting’) of Beckenbach’s sophism. . . .</p>
<p>I [Sorensen] agree with Beckenbach that some numbers are interesting. . . . But I also think that there are infinitely many uninteresting integers. Since the dull integers must start somewhere, there must be a first one &#8212; even if the vagueness of `dull’ makes it impossible to specify which it is. . . .</p>
<p>Beckenbach bases premise B on the fact that any instance of D will imply E:</p>
<p>D. It is interesting that n is the first uninteresting integer.</p>
<p>E. Therefore, n is an interesting integer.</p>
<p>My counter-explanation of the unsoundness is that the inference is invalid. . . .</p></blockquote>
<p>Sorensen&#8217;s key point is that a number can be embedded in an interesting statement without itself being interesting. Suppose, for example, we say that each of the integers from 0 through 20 are interesting, as are 23 and 24, but that 21 and 22 are dull (with the terms &#8220;dull&#8221; and &#8220;interesting&#8221; depending on context; the set of numbers that are interesting to a sports fan could differ from the set of numbers that are interesting to a mathematician; and also conditional on the level of focus, as the harder you look the more likely it is that you can find something interesting; but that doesn&#8217;t affect the paradox, it&#8217;s just a matter of definition). It&#8217;s arguably an interesting statement that 21 is the smallest dull positive integer, but that doesn&#8217;t make 21 itself interesting.</p>
<p>Sorensen gives several examples of that sort of thing:</p>
<blockquote><p>An uninteresting fact can embed an interesting fact. (For instance, it is interesting that the coastline of Norway is longer than the coastline of the United States but it is not interesting that this fact is interesting.) The case for D centers on the dual of this embedding principle: An interesting fact can incorporate a dull fact.</p>
<p>Indeed, it can be interesting that a fact is dull . . . `873 is the difference between the squares of two consecutive integers’ looks interesting. But actually this fact is not interesting; any odd number greater than 1 is the difference between the squares of two consecutive integers. The dullness of `873 is the difference between the squares of two consecutive integers’ is interesting because this dullness is explained by an interesting generality.</p>
<p>Undistinctiveness is just one genre of instructive dullness. The monotony of `The decimal expansion of 1/9 is .111….’ is a sign that it is a non-terminating fraction. The enervating patternlessness of the decimal expansion of π is a sign that it is a transcendental number.</p>
<p>In an early example of computer program, Alan Turing analyzed chess into a sequence subtasks. The more menial he made the sub procedures, the more interest he added to the overall effect. Turing’s chess programs breathed new life into homuncular models of psychological processes. . . .</p>
<p>In defense of conceptual analysis, I say the dullness of an identity statement often promotes the interest of the analysis. Consider contested identities. Students at first deny 1 = .999…. The teacher then points out that 1/3 = .333…. and 1/3 x 3 = 1. This conjunction of trivial truths makes most students regard 1 and .999…. as alternate numeric representations of the same number. They stop viewing 1 = .999…. as a near miss and start regarding it as trivially true. In the final analysis, the interesting fact is not that 1 = .999…. Just the opposite! The interesting fact is that `1 = .999….’ is dull.</p></blockquote>
<p>Or, for an even simpler example, you can write an article with about the color green, without that article itself being green.</p>
<p>The questions become more subtle when we move between mathematics and the social world (including the very use of base 10 as a social convention):</p>
<blockquote><p>A dull number can be denoted by an interesting numeral. In hexadecimal (base 16), 570005 is denoted by DEAD. . . .</p>
<p>Once we become sensitized to the distinction between using an expression and merely mentioning it, we become more discriminating about the means by which a number can inherit interest from a fact. . . .</p>
<p>In mathematics, inheritance is restricted to internal relations. The interest of `92 is the number of different arrangements of 8 non-attacking queens on an 8 x 8 chessboard’ is assigned to chess rather than 92 because chess is an alien relatum.</p>
<p>People relate interestingly to numbers but the interest of these relationships attaches to people. The interest of `The grandmaster Bobby Fisher died at 64, the number of squares on a chessboard’ attaches to Fisher, not 64.</p>
<p>Plutarch remarks that &#8220;The Pythagoreans also have a horror for the number 17, for 17 lies exactly halfway between 16, which is a square, and the number 18, which is the double of a square, these two, 16 and 18, being the only two numbers representing areas for which the perimeter equals the area&#8221;. This is an interesting fact about <em>Pythagoreans</em>. Their horror does not add to the interest of 17 (though 17 may accrue interest from the mathematical relationship that troubled the Pythagoreans).</p>
<p>In 1866, sixteen year old B. Nicolò I. Paganini found the small amicable pair (1184, 1210). . . . Paganini’s amicable pair is interesting in that it partly answers `Which are the amicable pairs?’. But it is more interesting as evidence for the <em>psychological</em> question `How reliable were the great mathematicians?’. Mathematicians are reluctant to credit a pair of numbers with the interest that attaches to <em>contingent</em> facts about it.</p>
<p>Another obstacle to crediting interest to Paganini’s numbers 1184 and 1210 is that they are interesting <em>as a pair</em>. Interest in a pair need not pass down to the individuals comprising the pair (just as a husband and wife can each be bores and yet be interesting as a couple).</p></blockquote>
<p>Also this:</p>
<blockquote><p>The robustness of interesting dullness is manifested by the sheer volume of commentary on boredom. Philosophers marvel at the power of this motivational vacuum. . . . Social scientists agree that just as there can be sober studies of inebriation, there can be interesting studies of tedium, repetition, and apathy. . . .</p>
<p>When astronomers explain away coincidences with an identity hypothesis, loss of wonder is experienced as insight. Demystification is a sign of explanatory progress.</p></blockquote>
<p><strong>The balance of boredom</strong></p>
<p>Sorensen&#8217;s article is full of striking insights, for example:</p>
<blockquote><p>Satiation differs from boredom in that you can exit. The gorged gourmand just leaves the restaurant. But the dishwasher is obliged to stay. World weary, the dishwasher can only escape into daydreams and diversions. Boredom correlates with understanding. So there is some temptation to compress Heidegger into two lines: To understand everything is to be bored by everything. So everything is boring.</p>
<p>But boredom can also be produced by incomprehension or a slight distraction (too small to be recognized as the true cause of one’s inability to focus). The laggard is too far behind to make sense of the lesson. The prodigy is too far ahead to find the lesson stimulating. The interested student lies in between, challenged but not overwhelmed.</p>
<p>Prudent students monitor their boredom to check whether they have deviated from this balance. The vain misconstrue the boredom of incompetence as the boredom of mastery.</p></blockquote>
<p><strong>Relation to the philosophy or sociology of science</strong></p>
<p>The discussion in the second half of Sorensen&#8217;s article reminds me of some things we&#8217;ve talked about regarding bad science:</p>
<p>1. My false theorem. I proved a false theorem once! <a href="http://stat.columbia.edu/~gelman/research/published/GelmanSpeed.pdf">Here&#8217;s the original article</a> (Andrew Gelman and T. P. Speed (1993), Characterizing a joint probability distribution by conditionals, Journal of the Royal Statistical Society B 55, 185-188), and <a href="http://stat.columbia.edu/~gelman/research/published/GelmanSpeedCorrection.pdf">here&#8217;s the correction notice</a>, from 1999. Embarrassingly for us, the falseness of our claim was shown by a one-line counterexample. Here&#8217;s the relevance to the present discussion: I always referred to this as our &#8220;false theorem,&#8221; but then someone pointed out that it&#8217;s not a theorem if it&#8217;s false! So what word to use? &#8220;Conjecture&#8221; doesn&#8217;t seem quite right, because we were offering it as a theorem, not a hypothesis. &#8220;Claim&#8221; is better, but that doesn&#8217;t convey that we were not just saying that a particular statement was true; we were claiming we&#8217;d proved it. I think this difficulty in explaining is &#8220;real,&#8221; not just a matter of the lack of a good word in English for &#8220;claimed proof.&#8221;</p>
<p>2. Evidence vs. truth. We&#8217;ve talked about this many times, for example <a href="http://stat.columbia.edu/~gelman/research/published/AssessingEvidence.pdf">here</a>, <a href="https://statmodeling.stat.columbia.edu/2018/04/06/important-distinction-truth-evidence/">here</a>, and <a href="https://statmodeling.stat.columbia.edu/2024/12/22/stanford-medical-school-professor-misrepresents-what-i-wrote-but-i-kind-of-understand-where-hes-coming-from/">here</a>. I think this is a big issue with problematic science, that researchers will make a claim that might well be true, but they don&#8217;t offer good evidence. When a scientific paper P is published claiming to demonstrate statement X, what the paper is really claiming is not &#8220;X is true&#8221; but rather &#8220;P contains strong evidence in favor of the truth of X.&#8221; When an outsider (such as me!) criticizes the paper&#8217;s &#8220;methods,&#8221; we&#8217;re typically arguing against that second claim, i.e. we&#8217;re saying that P <em>does not</em> contain strong evidence in favor of the truth of X. The original authors of the paper will typically respond with some version of, &#8220;We believe that X is true,&#8221; which might be fine, but I think it impedes the discussion for them to not first accept that P does not contain strong evidence in favor of the truth of X&#8211;or at least to address the outsider&#8217;s criticism on that level.</p>
<p>3. <a href="http://stat.columbia.edu/~gelman/research/published/ethics25.pdf">Big if true</a>. This comes up a lot, for example studies of extra-sensory perception, or the claim that women are three times more likely to wear red or pink clothing during certain times of the month, or claims that subliminal messages can cause huge opinion swings, or claims of a stolen election in 2020, or various other topics we&#8217;ve covered in this space over the years: these claims are implausible on their face, and a careful look at the published evidence offered in their support do not change this assessment. But, if they were true, they&#8217;d be interesting! Thus, as Sorensen discusses, these are statements whose interestingness depends on their truth value.</p>
<p><strong>P.S.</strong> I just wrote this post this morning. The next slot on the schedule is in May, but I bumped today&#8217;s scheduled post and stuck in this one instead, because the topic is so interesting (to me). Really.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/30/a-very-interesting-discussion-by-roy-sorenson-of-the-interesting-number-paradox/feed/</wfw:commentRss>
			<slash:comments>29</slash:comments>
		
		
			</item>
		<item>
		<title>Sorry, NYT, but, yes, &#8220;Equidistant Letter Sequences in the Book of Genesis&#8221; was junk science</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/29/sorry-nyt-but-yes-equidistant-letter-sequences-in-the-book-of-genesis-was-junk-science/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/29/sorry-nyt-but-yes-equidistant-letter-sequences-in-the-book-of-genesis-was-junk-science/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Sun, 29 Dec 2024 14:13:30 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=50986</guid>

					<description><![CDATA[From the New York Times: It sounds like a headline ripped from a supermarket tabloid: In 1994, three Israeli researchers claimed to have found a secret code embedded in Genesis, the first book of the Old Testament. But this wasn’t &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/29/sorry-nyt-but-yes-equidistant-letter-sequences-in-the-book-of-genesis-was-junk-science/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><a href="https://www.nytimes.com/2024/08/30/science/eliyahu-rips-dead.html">From the New York Times</a>:</p>
<blockquote><p>It sounds like a headline ripped from a supermarket tabloid: In 1994, three Israeli researchers claimed to have found a secret code embedded in Genesis, the first book of the Old Testament.</p>
<p>But this wasn’t junk science.</p></blockquote>
<p>Ahhhh . . . but it was!</p>
<p>The Times continues:</p>
<blockquote><p>The paper in which they revealed their findings appeared in an esteemed, peer-reviewed journal. And the academic reputations of the three authors — Eliyahu Rips, Yoav Rosenberg and Doron Witztum — were unimpeachable, especially that of Dr. Rips.</p></blockquote>
<p>The reporter&#8217;s mistake was to think that, just cos something&#8217;s authored by professors at a legitimate journal and published in a reputable journal, that it can&#8217;t be junk science.</p>
<p>An understandable mistake to make in 1994, but hard to support nowadays.  <a href="https://statmodeling.stat.columbia.edu/2023/08/31/the-variation-ignoring-junk-science-thats-promoted-by-association-for-psychological-science-and-related-academic-celebrities-its-like-a-poker-player-thinking-okay-if-push-all/">Lucky golf ball</a>, anyone?</p>
<p><strong>P.S.</strong>  I remember talking with Dave Krantz a couple years after that Bible code paper came out.  Some of his colleagues were really bothered by it and went to the trouble of <a href="https://www.math.toronto.edu/~drorbn/Codes/Chance.pdf">figuring out and explaining</a> what was going on.  Dave was kind of irritated that they were wasting any time on it at all, and I just thought the whole thing was a big joke.</p>
<p>In retrospect I think Dave and I should&#8217;ve shown more respect to the debunkers.  The Bible code paper was an early example of junk science leveraging the authority of the academic community to get publicity.</p>
<p>The junk-science-to-journal-to-NPR/Ted pipeline must have seemed like a great deal for everyone at the time:  more publicity for researchers, more credibility for unconventional scientific claims, more fun feature stories for the news media.  And for awhile it seemed to be going just fine, turning Malcolm Gladwell into a New Yorker celebrity, powering the Freakonomics franchise, and providing raw material for Jeffrey Epstein&#8217;s Edge Foundation.  The Pizzagate and Shreddergate researchers became wildly successful, and even fringe players such as the beauty-and-sex-ratio guy were able to land book contracts.  Formerly obscure law professors got to mingle with Henry Kissinger!  It was boom time in This Week in Psychological Science.</p>
<p>Eventually the weight of the junk science <a href="https://statmodeling.stat.columbia.edu/2016/09/21/what-has-happened-down-here-is-the-winds-have-changed/">overwhelmed the system</a>, and ultimately I think that the whole thing was a bit of a deal with a devil:  In the short term, academic social science got lots of publicity, and a few well-placed and credulous or unscrupulous professors became media stars.  In the medium term, academic journals justly lost much of their authority.  In the longer term, who knows.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/29/sorry-nyt-but-yes-equidistant-letter-sequences-in-the-book-of-genesis-was-junk-science/feed/</wfw:commentRss>
			<slash:comments>20</slash:comments>
		
		
			</item>
		<item>
		<title>Bayesian inference (and mathematical reasoning more generally) isn&#8217;t just about getting the answer; it&#8217;s also about clarifying the mapping from assumptions to inference to decision.</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/28/bayesian-inference-isnt-just-about-getting-the-answer-its-also-about-clarifying-the-mapping-from-assumptions-to-inference-to-decision/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/28/bayesian-inference-isnt-just-about-getting-the-answer-its-also-about-clarifying-the-mapping-from-assumptions-to-inference-to-decision/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Sat, 28 Dec 2024 14:21:39 +0000</pubDate>
				<category><![CDATA[Bayesian Statistics]]></category>
		<category><![CDATA[Decision Analysis]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=50447</guid>

					<description><![CDATA[Palko writes: I&#8217;m just an occasional Bayesian (and never an exo-biologist) so maybe I&#8217;m missing some subtleties here, but I believe educated estimates for the first probability vary widely with some close to 0 and some close to 1 with &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/28/bayesian-inference-isnt-just-about-getting-the-answer-its-also-about-clarifying-the-mapping-from-assumptions-to-inference-to-decision/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Palko writes:</p>
<blockquote><p>I&#8217;m just an occasional Bayesian (and never an exo-biologist) so maybe I&#8217;m missing some subtleties here, but I believe educated estimates for the first probability vary widely with some close to 0 and some close to 1 with no good sense of the distribution. Is there any point in applying the theorem at that point?  From <a href="https://www.wired.com/story/search-for-alien-life-biosignatures-gas-atmosphere/?utm_source=pocket-newtab-en-us">this Wired article</a>:</p>
<blockquote><p>If or when scientists detect a putative biosignature gas on a distant planet, they can use a formula called Bayes’ theorem to calculate the chance of life existing there based on three probabilities. Two have to do with biology. The first is the probability of life emerging on that planet given everything else that’s known about it. The second is the probability that, if there is life, it would create the biosignature we observe. Both factors carry significant uncertainties, according to the astrobiologists Cole Mathis of Arizona State University and Harrison Smith of the Earth-Life Science Institute of the Tokyo Institute of Technology, who explored this kind of reasoning in a paper last fall.</p></blockquote>
</blockquote>
<p>My reply:  I guess it&#8217;s fine to do the calculation, if only to make it clear how dependent it is on <a href="https://statmodeling.stat.columbia.edu/2019/12/16/sumps-and-rigor/">assumps</a>.  Bayesian inference isn&#8217;t just about getting the answer; it&#8217;s also about clarifying the mapping from assumptions to inference to decision.</p>
<p>Come to think about it, that last paragraph remains true if you replace &#8220;Bayesian inference&#8221; with &#8220;Mathematics.&#8221;</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/28/bayesian-inference-isnt-just-about-getting-the-answer-its-also-about-clarifying-the-mapping-from-assumptions-to-inference-to-decision/feed/</wfw:commentRss>
			<slash:comments>16</slash:comments>
		
		
			</item>
		<item>
		<title>Announcing two new members of our blogging team . . .</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/27/announcing-two-new-members-of-our-blogging-team/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/27/announcing-two-new-members-of-our-blogging-team/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Fri, 27 Dec 2024 14:01:25 +0000</pubDate>
				<category><![CDATA[Literature]]></category>
		<category><![CDATA[Sports]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=50980</guid>

					<description><![CDATA[. . . Nick Hornby and David Roche! Just kidding. Really, though, these guys are great and should be blogging for us. Nick Hornby you know about. He&#8217;s the author of High Fidelity, the book that Phil once said that &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/27/announcing-two-new-members-of-our-blogging-team/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>. . . Nick Hornby and David Roche!</p>
<p>Just kidding.</p>
<p>Really, though, these guys are great and should be blogging for us.</p>
<p>Nick Hornby you know about.  He&#8217;s the author of High Fidelity, the book that Phil once  said that no woman should be allowed to read because it&#8217;s a cheat code for understanding guys.  I&#8217;ve been reading an old collection of his book columns for the Believer, and it&#8217;s so bloggy . . . it would fit in so well right here.  OK, he&#8217;s not gonna be blogging here; his monthly blog (I guess that&#8217;s too low a frequency of posts to be a &#8220;blog,&#8221; so let&#8217;s call it a &#8220;colummn&#8221;) is <a href="https://www.thebeliever.net/type/stuff-ive-been-reading/">here</a> . . . I guess I&#8217;ll put it in the <a href="https://statmodeling.stat.columbia.edu/blogs-i-read/">Blogs We Read</a> page.</p>
<p>Hornby&#8217;s what <a href="https://statmodeling.stat.columbia.edu/2010/07/28/james_clueless/">I&#8217;ve called</a> a writer in the David Owen mode (that is, David Owen the American journalist, not David Owen the British politician): serious, earnest, somewhat intelligent but a bit of a blockhead. Which I mean in a good way. Not clever-clever or even clever, but he wants to get things right.  There&#8217;s kind of a paradox:  Hornby&#8217;s shtick is that he&#8217;s an unpretentious regular guy, writing about unpretentious regular guys&#8212;also, though, he&#8217;s a brilliant and successful author.  It&#8217;s a little different than John Updike, who wrote about his uncultured hero Rabbit, while maintaining the urbane &#8220;Updike&#8221; character for his public image.  Also different from Kingsley Amis, who basically tore himself apart in his attempt to be the regular guy befitting his literary and political ideology.  In his blogging columns, Hornby is pretty upfront about the tension between his literary-intellectual and regular-guy personas.</p>
<p>Our other new blogger (not really) is David Roche, who I&#8217;d never heard of until I read <a href="https://defector.com/how-david-roche-nearly-ran-his-dick-off-and-maybe-changed-ultrarunning-in-the-process">this interview of him</a> by Dennis Young.  Roche is a funny guy, also very analytical, would be a good blogger for us.  Like Hornby, he also has <a href="https://www.trailrunnermag.com/byline/david-roche/">a column</a>, but it&#8217;s in Trail Runner magazine, and it&#8217;s full of articles like &#8220;New Study Shows Strategically Reduced Carbohydrate Intake Does Not Improve Performance,&#8221; and I could care less about trail running.  That interview, though, it&#8217;s great.</p>
<p>Here&#8217;s a sample:</p>
<blockquote><p>A few studies have come out with athletes pushing 120 grams of carbs per hour, showing improved fatigue resistance late in events. But my mind was opened to how I could solve the Leadville equation during Stage 18 of the Tour de France. Victor Campenaerts won the stage unexpectedly, and after the stage, his sponsor Precision Fueling &#038; Hydration released his data (rare in cycling, where everyone holds secrets like they’re Gollum). He did 132 g/hr, pushing closer to 150 g/hr at times. I had never heard of anyone trying that and succeeding at such a high level. So I went for it, and I think I showed that it&#8217;s possible in running too.</p>
<p>Cycling is conducting the biggest uncontrolled performance experiment in the world. In ultras, the margins of human performance are not here yet. I was 1.6 percent ahead of Matt Carpenter, and I bet someone is going to be a few percent ahead of me. In cycling, everyone has (very roughly) the same power at baseline. When the margins are that narrow, the best training/fueling wins, and that&#8217;s one reason why doping has such a troubled history in that sport&#8212;it’s a confounding variable that fucks up a true understanding of human biking performance over longer time horizons.</p></blockquote>
<p>All this self-experimentation . . . he&#8217;s kind of like a sane <a href="https://statmodeling.stat.columbia.edu/2014/04/30/seth-roberts/">Seth Roberts</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/27/announcing-two-new-members-of-our-blogging-team/feed/</wfw:commentRss>
			<slash:comments>7</slash:comments>
		
		
			</item>
		<item>
		<title>Softmax is on the log, not the logit scale</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/26/those-are-unnormalized-log-probabilities-not-logits-in-your-neural-networks-final-layer/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/26/those-are-unnormalized-log-probabilities-not-logits-in-your-neural-networks-final-layer/#comments</comments>
		
		<dc:creator><![CDATA[Bob Carpenter]]></dc:creator>
		<pubDate>Thu, 26 Dec 2024 20:00:52 +0000</pubDate>
				<category><![CDATA[Miscellaneous Statistics]]></category>
		<category><![CDATA[Stan]]></category>
		<category><![CDATA[Statistical Computing]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=50149</guid>

					<description><![CDATA[Bad Stan naming I realized recently that we followed the confusing terminological convention of ML in our description of Stan&#8217;s categorical_logit function. In Stan, if there&#8217;s a suffix to a distribution, it describes the scale of one or more of &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/26/those-are-unnormalized-log-probabilities-not-logits-in-your-neural-networks-final-layer/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><b>Bad Stan naming</b></p>
<p>I realized recently that we followed the confusing terminological convention of ML in our description of Stan&#8217;s <code>categorical_logit</code> function.  In Stan, if there&#8217;s a suffix to a distribution, it describes the scale of one or more of the parameters.  For example, </p>
<p><code>poisson_log(y | u) == poisson(y | exp(u))</code>.  </p>
<p>So when we write <code>categorical(y | p)</code> we take <code>p</code> to be a simplex (sequence of finite, non-negative values that sum to 1).  So it would make sense that <code>categorical_logit(y | logit(p))</code> would be equivalent, where <code>logit(p) = log(p / (1 - p))</code>. But that&#8217;s not how it works in Stan.  Instead,</p>
<p><code>caetgorical_logit(y | u) = categorical(y | softmax(u))</code>. </p>
<p>We made the same mistake everyone on ML makes in their variable naming! We call the <code>u</code> here &#8220;logits&#8221;, when in fact they&#8217;re (unnormalized [see below]) log probabilities.  This is probably due to the fact that if <code>u</code> is a regression, then the resulting system is called &#8220;multinomial logistic regression.&#8221;  </p>
<p><b>Example</b></p>
<p>The softmax function is defined by <code>softmax(u) = exp(u) / sum(exp(u))</code>.  When used like this, the arguments to softmax are log probabilities, not logit probabilities.  Here&#8217;s a little snippet of Python to illustrate (the style sheet is adding the extra space, not me, and I don&#8217;t want to fix it manually in this post with a hack because it&#8217;ll mess up the page if the style sheet is ever fixed).  </p>
<p><code><br />
>>> p = np.asarray([0.2, 0.5, 0.3])</p>
<p>>>> def logit(p): return np.log(p / (1 - p))<br />
... </p>
<p>>>> logit_p = logit(p)</p>
<p>>>> log_p = np.log(p)</p>
<p>>>> sp.special.softmax(logit_p)<br />
array([0.14893617, 0.59574468, 0.25531915])</p>
<p>>>> sp.special.softmax(log_p)<br />
array([0.2, 0.5, 0.3])<br />
</code></p>
<p>This shows that for the round trip probabilities through softmax, the appropriate operation is the natural logarithm, not the logit function.</p>
<p><b>Origin of the confusion</b></p>
<p>So where did this confusion come from?  Let&#8217;s look at a standard binary logistic regression.  There we take</p>
<p><code><br />
p(y | alpha, beta, x) = bernoulli(y | inv_logit(alpha + beta * x))<br />
</code></p>
<p>where </p>
<p><code>inv_logit(v) = exp(v) / (1 + exp(v))</code>.  </p>
<p>Writing inverse logit this way suggests how to write a logistic regression with a categorical distribution and softmax.</p>
<p><code><br />
p(y | alpha, beta, x) = categorical(y | softmax([0, alpha + beta * x]))<br />
</code></p>
<p>that&#8217;s because </p>
<p><code>softmax([0, alpha + beta * x])<br />
    = [exp(0), exp(alpha + beta * x)] / (exp(0) + exp(alpha + beta * x))<br />
    = [1, exp(alpha + beta(x)] / (1 + exp(alpha + beta * x))<br />
    = [1 / (1 + exp(alpha + beta * x),  exp(alpha + beta * x) / (1 + exp(alpha + beta * x)]<br />
    = [1 - inv_logit(alpha + beta * x),  inv_logit(alpha + beta * x)],<br />
</code></p>
<p>This derivation shows that the probability of the categorical in this formulation returning 1 is <code>inv_logit(alpha + beta * x)</code>.  But this connection falls apart in the multinomial case when there are more than two outcomes.</p>
<p>In traditional frequentist <code>K</code> outcome multinomial logistic regressions, the first input to softmax is pinned to 0 for identifiability just as in the binary case.  </p>
<p><code>softmax([0, u[2], ..., u[K1])<br />
    = [exp(0), exp(u[2]), ..., exp(u[K])] / (exp(0) + exp(u[2]) + ... + exp(u[K]))<br />
</code></p>
<p>This leads to asymmetry in the regression as we don&#8217;t have a regression for the first element.  What it does do is make softmax and log proper inverses.   If you reduce the choice to just the first category and some other category, then you get a standard binomial logistic regression again.  But you still can&#8217;t round trip the multinomial case with logit, because</p>
<p><code><br />
exp(u[2]) / (exp(0) + exp(u[2]) + ... + exp(u[K]))  !=  inv_logit(u[2])<br />
</code></p>
<p>To see that this is still not going to produce logits in the multinomial case, here&#8217;s some more Python.</p>
<p><code><br />
>>> log_p<br />
array([-1.60943791, -0.69314718, -1.2039728 ])</p>
<p>>>> log_p_zero = log_p - log_p[0]</p>
<p>>>> log_p_zero<br />
array([0.        , 0.91629073, 0.40546511])</p>
<p>>>> sp.special.softmax(log_p_zero)<br />
array([0.2, 0.5, 0.3])<br />
</code></p>
<p>So as you can see, softmax isn&#8217;t identified without pinning one of the values&#8212;we can add or subtract a constant from each element of the input and get the same value.  But this still doesn&#8217;t turn the inputs to softmax into logits.</p>
<p>>>> def inv_logit(v): 1 / (1 + exp(-v))<br />
&#8230; </p>
<p>>>> inv_logit(log_p_zero)<br />
array([0.5       , 0.71428571, 0.6       ])<br />
</code></p>
<p>So you can see that the input <code>0.91629073</code> is not the logit of the probability even when pinning a value to zero to identify.</p>
<p>P.S.  I really miss being able to write math on the blog and really hate that all my old posts with math no longer render. Maybe if Andrew reminds us why it went away, someone will have a suggestion on how to fix.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/26/those-are-unnormalized-log-probabilities-not-logits-in-your-neural-networks-final-layer/feed/</wfw:commentRss>
			<slash:comments>5</slash:comments>
		
		
			</item>
		<item>
		<title>Data manipulation in the world of long-distance swimming!</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/26/data-manipulation-in-the-world-of-long-distance-swimming/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/26/data-manipulation-in-the-world-of-long-distance-swimming/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Thu, 26 Dec 2024 14:04:27 +0000</pubDate>
				<category><![CDATA[Statistical Computing]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=50971</guid>

					<description><![CDATA[The story is here. It&#8217;s kind of like the Excel error in that famous economics paper, except that (a) it seems to have been on purpose, not an accident (according to the alleged perp, &#8220;The hacker has hacked into my &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/26/data-manipulation-in-the-world-of-long-distance-swimming/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><a href="https://defector.com/lawsuit-open-water-swimming-expert-rewrote-swimming-history-to-glorify-diana-nyad">The story is here</a>.  It&#8217;s kind of like <a href="https://statmodeling.stat.columbia.edu/2013/04/16/memo-to-reinhart-and-rogoff-i-think-its-best-to-admit-your-errors-and-go-on-from-there/">the Excel error</a> in that famous economics paper, except that (a) it seems to have been on purpose, not an accident (according to the alleged perp, &#8220;The hacker has hacked into my administrative account and is using my accounts to make changes&#8221;), and (b) it did not do any damage to the world economy, it just enabled a fictionalized story to be presented as fact in a Hollywood movie.</p>
<p>So, no biggie except for people involved in long-distance swimming&#8212;for them it&#8217;s a big deal.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/26/data-manipulation-in-the-world-of-long-distance-swimming/feed/</wfw:commentRss>
			<slash:comments>8</slash:comments>
		
		
			</item>
		<item>
		<title>How to cheat at Codenames; cheating at board games more generally</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/25/how-to-cheat-at-codenames-cheating-at-board-games-more-generally/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/25/how-to-cheat-at-codenames-cheating-at-board-games-more-generally/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Wed, 25 Dec 2024 14:36:02 +0000</pubDate>
				<category><![CDATA[Decision Analysis]]></category>
		<category><![CDATA[Sports]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=50954</guid>

					<description><![CDATA[This is a good post for Christmas Day, with all of you at home with your families playing board games. Dan Luu has an amusing post explaining how you can win Codenames by just memorizing the configurations of the 40 &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/25/how-to-cheat-at-codenames-cheating-at-board-games-more-generally/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><em>This is a good post for Christmas Day, with all of you at home with your families playing board games.</em></p>
<p>Dan Luu has <a href="https://danluu.com/codenames/">an amusing post</a> explaining how you can win Codenames by just memorizing the configurations of the 40 setup carts.  The basic strategy is to play your best until you can figure out the unique configuration, then you win.  The fun part is that if you&#8217;re playing against a team that hasn&#8217;t learned this memorization trick, then you can win even if you don&#8217;t guess any words yourself&#8212;you just take advantage of the config information that you get from their correct guesses (along with any wrong guesses that come up)!  If both teams have memorized the 40 cards, then you get to a new level of strategy.</p>
<p>As Luu says, no one would want to play Codenames in this way.  The whole point of the game is to guess the words; if you&#8217;re gonna do it by memorizing patterns, why play the game in the first place?  On the other hand, he also points out that once this information is there, you can&#8217;t un-see it.  So it&#8217;s a balance.</p>
<p>This comes up in the rules of Codenames itself:  you&#8217;re not allowed to give clues that suggest the position of the word on the grid, nor are you allowed to make faces or otherwise give clues as people are guessing.  It can be hard to avoid giving this information sometimes!</p>
<p>More generally, most games can be &#8220;cracked&#8221; through a backdoor approach in some way or another.  Here&#8217;s how Luu puts it:</p>
<blockquote><p>Personally, when I run into a side-channel attack in a game or a game that&#8217;s just totally busted if played to win . . . I think it makes sense to try to avoid &#8220;attacking&#8221; the game to the extent possible. I think this is sort of impossible to do perfectly in Codenames because people will form subconscious associations (I&#8217;ve noticed people guessing an extra word on the first turn just to mess around, which works more often than not — assuming they&#8217;re not cheating, and I believe they&#8217;re not cheating, the success rate strongly suggests the use some kind of side-channel information. That doesn&#8217;t necessarily have to be positional information from the cards, it could be something as simple as subconsciously noticing what the spymasters are intently looking at.</p>
<p><a href="https://mastodon.social/@danluu/110544419353766175">Dave Sirlin calls anyone who doesn&#8217;t take advantage of any legal possibility to win is a sucker (he derogatorily calls such people &#8220;scrubs&#8221;)</a> (he says that you should use cheats to win, like using maphacks in FPS games, as long as tournament organizers don&#8217;t ban the practice, and that tournaments should explicitly list what&#8217;s banned, avoiding generic &#8220;don&#8217;t do bad stuff&#8221; rules). I think people should play games however they find it fun and should find a group that likes playing games in the same way. If Dave finds it fun to memorize arbitrary info to win all of these games, he should do that. The reason I, as Dave Sirlin would put it, play like a scrub, for the kinds of games discussed here is because the games are generally badly broken if played seriously and I don&#8217;t personally find the ways in which they&#8217;re broken to be fun.</p></blockquote>
<p>It gets tricky sometimes, though.  Consider those goofy words that are in the Scrabble dictionary but aren&#8217;t really words, for example ef (&#8220;the letter F&#8221;) or po (&#8220;a chamber pot&#8221;).  These are not English words!  On the other hand, when you&#8217;re actually playing and you see an opportunity for ef or po or whatever, it&#8217;s hard to deny yourself the opportunity.  In that case, there&#8217;s an easy solution:  the rules allow the players to agree on any dictionary ahead of time, so no need to use the Scrabble dictionary.  On the other hand, this will annoy serious players.</p>
<p>There&#8217;s more of gray area with collusion, which can &#8220;break&#8221; almost any multiplayer game.  In poker, collusion is a form of cheating.  I don&#8217;t know how casinos or informal games monitor or enforce the rule against collusion, but you&#8217;re not supposed to do it.  You&#8217;re allowed to lie in poker but not to cheat.</p>
<p>But what about a game such as Monopoly or Risk where bargaining is part of the game?  Here&#8217;s a simple strategy in a 3-player game of Monopoly that will up your odds of winning from 1/3 to nearly 1/2:  Before the game begins, pick one of the other players and agree to flip a coin, after which the winner of the flip will devote all their effort to helping the other player win.  That&#8217;s easy enough to do:  just buy whatever property that comes up and sell to the other player for $1.  It won&#8217;t guarantee a win but it&#8217;s gotta take the win probability to very close to 100%.  Similarly with Risk.  Now, nobody&#8217;s gonna play this way because it&#8217;s no fun (except maybe once as a joke).  To put it another way, &#8220;winning a game of Monopoly or Risk&#8221; does not have much positive value in itself; the fun is in winning the game legitimately.  Again, though, there is a gray zone, and other players will rightly get annoyed if they see player A deliberately trying to help player B without there being a good reason in the context of a game.  In Risk, &#8220;I won&#8217;t attack you here if you don&#8217;t attack me there&#8221; is a legitimate strategy, but &#8220;I don&#8217;t attack you because I want to help you win&#8221; is not so cool.</p>
<p>A few years ago I was playing a lot of online chess, and one thing I noticed is that some players would set up opening traps:  clearly unsound sequences of moves that would get them a win if their opponents played naively and hadn&#8217;t seen the trick before.  My thought was:  Why do that?  Winning against a stranger using a trap, what&#8217;s the point of that?  Upon reflection, though, I decided to not be so bothered by this.  If you try to spring a trap, then the fun part is when the trap fails and then you have to get out of a bad position of your own devising.  So, all good.</p>
<p>Years ago I read the book Thursday Night Poker by Peter Steiner.  One thing Steiner discusses is that in a casual game you can often do just fine by playing really tight, a strategy that won&#8217;t work against good players but can make you steady money if some of the people at the table are just playing for fun.  As Steiner says, though, most of us are not playing in a friendly poker game with the goal of maximizing our dollars.  We&#8217;re playing poker for fun, and &#8220;action&#8221;&#8212;getting involved in hands, making betting decisions, going up against the other players&#8212;is where the fun is at.  No poker player would be a &#8220;scrub&#8221;&#8212;you&#8217;ll always take advantage of any legal way to win, it&#8217;s not like you&#8217;d ignore relevant information that someone reveals&#8212;but, even in poker, winning is not the only goal.</p>
<p>All of this is kind of obvious, but as Luu discusses, sometimes it needs to be pointed out, to push against naive models of the world.  Also, the bit about the Codenames cards is cool&#8212;I&#8217;d never thought about that!</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/25/how-to-cheat-at-codenames-cheating-at-board-games-more-generally/feed/</wfw:commentRss>
			<slash:comments>11</slash:comments>
		
		
			</item>
		<item>
		<title>Addressing legitimate counterarguments in a scientific review:  The challenge of being an insider</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/24/addressing-legitimate-counterarguments-in-a-scientific-review-the-challenge-of-being-an-insider/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/24/addressing-legitimate-counterarguments-in-a-scientific-review-the-challenge-of-being-an-insider/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Tue, 24 Dec 2024 14:40:07 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Multilevel Modeling]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=50963</guid>

					<description><![CDATA[Review articles can be written by outsiders or insiders. From the outside, it&#8217;s easier. You assess the evidence and draw your conclusions. For example, this is what I did when summarizing the research on ballot-order effects and addressing the question &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/24/addressing-legitimate-counterarguments-in-a-scientific-review-the-challenge-of-being-an-insider/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Review articles can be written by outsiders or insiders.</p>
<p>From the outside, it&#8217;s easier.  You assess the evidence and draw your conclusions.  For example, this is <a href="https://statmodeling.stat.columbia.edu/2017/02/27/name-first-doubtful/">what I did</a> when summarizing the research on ballot-order effects and addressing the question of whether Donald Trump won the 2016 election because his name came first on the ballot in key states; see pages 240-242 of <a href="https://stat.columbia.edu/~gelman/active-statistics/">Active Statistics</a> for the full story.  Or when my colleagues and I <a href="http://stat.columbia.edu/~gelman/research/published/PNAS_letter_Nudge-meta-analysis.pdf">wrote about</a> &#8220;nudge&#8221; interventions.  We could be right, we could be wrong, but in any case the job is clear enough.</p>
<p>Writing a review is more difficult from the inside.  For example, consider the meta-analysis of nudge that was written by some nudge researchers.  <a href="https://statmodeling.stat.columbia.edu/2022/01/10/the-real-problem-of-that-nudge-meta-analysis-is-not-that-it-include-12-papers-by-noted-fraudsters-its-the-gigo-of-it-all/">It had big problems</a>:  garbage in, garbage out.  It&#8217;s hard to step back and examine the evidence, if you&#8217;re part of the story.  For another example, <a href="https://statmodeling.stat.columbia.edu/2024/08/20/when-is-it-appropriate-to-give-a-one-sided-perspective-implicit-association-test-example/">we recently discussed</a> a review of the controversial Implicit Association Test that was flawed in only presenting part of the story, not even acknowledging that there was a controversy.</p>
<p>What to do if you&#8217;re an insider and you want to write a review that includes some of your own work?</p>
<p>I don&#8217;t think you should just give up.  As an insider, you have a special perspective, and it makes sense that you&#8217;ll want to review the evidence.  At the same time, you have to avoid the natural inclination to try to present too much of a coherent story.  In real life, the evidence doesn&#8217;t always all go in the same direction!</p>
<p>My recommendation, for the insider writing a review article, is not to try to debate or shoot down such counterarguments but rather to acknowledge the disagreement and fit it into your larger story.</p>
<p>Here&#8217;s an example where we did exactly that.  Our article is called <a href="http://stat.columbia.edu/~gelman/research/published/reconciling_millennium.pdf">Reconciling Evaluations of the Millennium Villages Project</a>, and it begins:</p>
<blockquote><p>The Millennium Villages Project was an integrated rural development program carried out for a decade in 10 clusters of villages in sub-Saharan Africa starting in 2005, and in a few other sites for shorter durations. An evaluation of the 10 main sites compared to retrospectively chosen control sites estimated positive effects on a range of economic, social, and health outcomes (Mitchell et al. 2018). More recently, an outside group performed a prospective controlled (but also nonrandomized) evaluation of one of the shorter-duration sites and reported smaller or null results (Masset et al. 2020). Although these two conclusions seem contradictory, the differences can be explained by the fact that Mitchell et al. studied 10 sites where the project was implemented for 10 years, and Masset et al. studied one site with a program lasting less than 5 years, as well as differences in inference and framing. Insights from both evaluations should be valuable in considering future development efforts of this sort. Both studies are consistent with a larger picture of positive average impacts (compared to untreated villages) across a broad range of outcomes, but with effects varying across sites or requiring an adequate duration for impacts to be manifested.</p></blockquote>
<p>&#8220;Mitchell et al. 2018&#8221; was us!  I worked with the Millennium Villages Project to help conduct a retrospective evaluation, which yielded positive estimated effects.  My take was that the project worked well.  That&#8217;s what I believe, but I&#8217;m an insider.  Masset et al. 2020 was an outside team that did a different analysis on different data and reported null results.  The purpose of our review article was to understand how these two studies of the same program came to such different conclusions.  In writing this paper we had to walk a fine line:  we&#8217;re trying our best to assess the past work objectively, but one of the papers was ours.  The key here is that we presented the disagreement&#8212;we did not pretend the dissenting article did not exist, nor did we dismiss it&#8212;; rather, we incorporated it into a larger understanding.  I&#8217;m not saying that this new paper of ours was perfect, nor that it was influential&#8212;indeed, according to Google scholar it has exactly zero citations, really a low payoff given all the work we put into it&#8212;but I still think of it as a model for how to present and assess conflicting evidence from the inside.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/24/addressing-legitimate-counterarguments-in-a-scientific-review-the-challenge-of-being-an-insider/feed/</wfw:commentRss>
			<slash:comments>4</slash:comments>
		
		
			</item>
		<item>
		<title>The marginalization or Jeffreys-Lindley paradox:  it&#8217;s already been resolved.</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/23/the-marginalization-or-jeffreys-lindley-paradox-its-already-been-resolved/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/23/the-marginalization-or-jeffreys-lindley-paradox-its-already-been-resolved/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Mon, 23 Dec 2024 14:48:12 +0000</pubDate>
				<category><![CDATA[Bayesian Statistics]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51504</guid>

					<description><![CDATA[Nicole Wang writes: I am a PhD student in statistics starting this year. I read the Dawid et al. (1973) marginalization paradoxes paper. I found several approaches claiming their “resolution” to this paradox. I am very curious whether you think &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/23/the-marginalization-or-jeffreys-lindley-paradox-its-already-been-resolved/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Nicole Wang writes:</p>
<blockquote><p>I am a PhD student in statistics starting this year.  I read the Dawid et al. (1973) marginalization paradoxes paper. I found several approaches claiming their “resolution” to this paradox. I am very curious whether you think there is a complete resolution to this paradox so far. If there is one, I sincerely wonder what it is. If not, which approach(s) do you think it(they) is(are) closer to the resolution?</p></blockquote>
<p>I replied by pointing to these two posts:<br />
<a href="https://statmodeling.stat.columbia.edu/2013/09/17/christian-robert-on-the-jeffreys-lindley-paradox-more-generally-its-good-news-when-philosophical-arguments-can-be-transformed-into-technical-modeling-issues/">Christian Robert on the Jeffreys-Lindley paradox; more generally, it’s good news when philosophical arguments can be transformed into technical modeling issues</a><br />
and<br />
<a href="https://statmodeling.stat.columbia.edu/2013/03/11/my-problem-with-the-lindley-paradox/">My problem with the Lindley paradox</a>.<br />
It&#8217;s not a topic I&#8217;ve thought about much lately because to me it&#8217;s already been resolved.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/23/the-marginalization-or-jeffreys-lindley-paradox-its-already-been-resolved/feed/</wfw:commentRss>
			<slash:comments>3</slash:comments>
		
		
			</item>
		<item>
		<title>Stanford medical school professor misrepresents what I wrote (but I kind of understand where he&#8217;s coming from)</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/22/stanford-medical-school-professor-misrepresents-what-i-wrote-but-i-kind-of-understand-where-hes-coming-from/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/22/stanford-medical-school-professor-misrepresents-what-i-wrote-but-i-kind-of-understand-where-hes-coming-from/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Sun, 22 Dec 2024 17:59:12 +0000</pubDate>
				<category><![CDATA[Bayesian Statistics]]></category>
		<category><![CDATA[Public Health]]></category>
		<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51502</guid>

					<description><![CDATA[This story is kinda complicated. It&#8217;s simple, but it&#8217;s complicated. The simple part is the basic story, which goes something like this: &#8211; In 2020, a study was done at Stanford&#8211;a survey of covid exposure&#8211;that I publicly criticized: I wrote &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/22/stanford-medical-school-professor-misrepresents-what-i-wrote-but-i-kind-of-understand-where-hes-coming-from/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>This story is kinda complicated.  It&#8217;s simple, but it&#8217;s complicated.</p>
<p><em>The simple part</em> is the basic story, which goes something like this:</p>
<p>&#8211; In 2020, a study was done at Stanford&#8211;a survey of covid exposure&#8211;that I publicly criticized:  I wrote that this study had statistical problems, that its analysis did not adjust for uncertainty in the false positive rate of the test they were using, and that they did something wrong by not making use of the statistical expertise that is readily available at Stanford.</p>
<p>&#8211; In 2023, one of the people involved in that study wrote a long post about that study and its aftermath. In that post, my comments from 2020 on that study were misrepresented&#8211;hence the title of the present post.</p>
<p>&#8211; Just last week someone pointed me to the 2023 post.  I was unhappy to have been misrepresented so I emailed the author of the post.  He didn&#8217;t respond&#8211;I don&#8217;t take that personally:  he&#8217;s very busy, the holiday season is coming, the post came out over a year ago (I just only happened to hear about it the other day), and my complaint concerns only one small paragraph in a long post&#8211;so, just to correct the record, I&#8217;m posting this here.</p>
<p><em>The more complicated</em>, and interesting, part, involves the distinction between evidence and truth.  It&#8217;s something we&#8217;ve talked about before&#8211;indeed, I published <a href="http://stat.columbia.edu/~gelman/research/published/AssessingEvidence.pdf">a short article on the topic</a> back in 2020 using this very example!&#8211;and here it has come again, so here goes:</p>
<p><img decoding="async" src="https://statmodeling.stat.columbia.edu/wp-content/uploads/2024/12/Screenshot-2024-12-22-at-11.49.15-1024x642.png" alt="" width="350" /></p>
<p>And now, the details:</p>
<p>Stanford professor Jay Bhattacharya <a href="https://www.jospi.org/article/88046-dr-jay-bhattacharya-reveals-stanford-university-s-attempts-to-derail-covid-studies">wrote about</a> a covid study from 2020 that he was involved in, which attracted some skepticism at the time:</p>
<blockquote><p>Some serious statisticians also weighed in with negative reviews. Columbia University’s Andrew Gelman posted a hyperbolic blog that we should apologize for releasing the study. He incorrectly thought we had not accounted for the possibility of false positives. He later recanted that harsh criticism but wanted us to use an alternative method of characterizing the uncertainty around our estimates.</p></blockquote>
<p>On the plus side, I appreciate that he characterizes me as a serious statistician.  He also called my post &#8220;hyperbolic.&#8221;  He doesn&#8217;t actually link to it, so I&#8217;ll <a href="https://statmodeling.stat.columbia.edu/2020/04/19/fatal-flaws-in-stanford-study-of-coronavirus-prevalence/">give the link here</a> so you can make your own judgment.  The title of that post is, &#8220;Concerns with that Stanford study of coronavirus prevalence,&#8221; and I don&#8217;t think it&#8217;s hyperbolic at all!  But that&#8217;s just a matter of opinion on Bhattacharya&#8217;s part so I can&#8217;t say it&#8217;s wrong.</p>
<p>He does have two specific statements there that <em>are</em> wrong, however:</p>
<p><strong>1.</strong>  It&#8217;s not true that I &#8220;incorrectly thought their study had not accounted for the possibility of false positives.&#8221;  In my post, I explicitly recognized that their analysis accounted for the possibility of false positives.  What I wrote is that they were &#8220;focusing on the point estimates of specificity.&#8221;  Specificity = 1 &#8211; false positive rate.  What I wrote is that they <em>didn&#8217;t properly account for uncertainty</em> in the false positive rate.  I did not say they had not accounted for the possibility of false positives.</p>
<p><strong>2.</strong>  I never &#8220;recanted that harsh criticism.&#8221;  What I wrote in my post is that their article &#8220;does not provide strong evidence that the rate of people in Santa Clara county exposed by that date was as high as claimed.&#8221;  But I also wrote, &#8220;I’m not saying that the claims in the above-linked paper are wrong . . . The Bendavid et al. study is problematic if it is taken as strong evidence for those particular estimates, but it’s valuable if it’s considered as one piece of information that’s part of a big picture that remains uncertain.&#8221;  And I clarified, &#8220;When I wrote that the authors of the article owe us all an apology, I didn’t mean they owed us an apology for doing the study, I meant they owed us an apology for avoidable errors in the statistical analysis that led to overconfident claims. But, again, let’s not make the opposite mistake of using uncertainty as a way to affirm a null hypothesis.&#8221;</p>
<p>I do not see this as a &#8220;recanting,&#8221; nor did I recant at any later time, but I&#8217;m fine if Bhattacharya or anyone else wants to quote me directly to make clear that at no time did I ever say that their substantive claims were false; I only said that the data offered in that study did not supply strong evidence.</p>
<p>This bothers me.  I&#8217;d hate for people to think that I&#8217;d incorrectly thought they had not accounted for the possibility of false positives, or that I&#8217;d recanted my criticism.  Again I emphasize that my criticism was statistical and involved specification of uncertainty; it was not a claim on my part that the percentage of people who&#8217;d been exposed to Covid was X, Y, or Z.</p>
<p>The good news is that this post from Bhattacharya appeared in 2023 and I only heard about it just the other day, so I guess it did not get wide circulation.  Maybe more people will see this correction than the original post!  In any case I&#8217;m glad to have the opportunity to correct the record.</p>
<p><strong>Stanford contagion</strong></p>
<p>I have no problem with Bhattacharya making arguments about covid epidemiology and policy&#8211;two topics he&#8217;s thought a lot about.  It&#8217;s completely reasonable for him and his colleagues to say that their original study was inconclusive but that it was consistent with their larger message.</p>
<p>Bhattacharya also writes:</p>
<blockquote><p>In the end, Stanford’s leadership undermined public and scientific confidence in the results of the Santa Clara study. Given this history, members of the public could be forgiven if they wonder whether any Stanford research can be trusted.</p></blockquote>
<p>He doesn&#8217;t fully follow up on this point, but I think he&#8217;s right.</p>
<p>A few weeks before the above-discussed covid study came out, Stanford got some press when law professor Richard Epstein published something through Stanford&#8217;s Hoover Institution predicting that U.S. covid deaths would max out at 500, a prediction he later updated to 5000 (<a href="https://rexdouglass.github.io/TIGR/Douglass_2020_How_To_Be_Curious_Instead_of_Contrarian_About_Covid19.nb.html">see here</a> for details).  I&#8217;ve never met Epstein or corresponded with him, but he comes off as quite the asshole, having said <a href="https://www.newyorker.com/news/q-and-a/the-contrarian-coronavirus-theory-that-informed-the-trump-administration">this</a> to a magazine interviewer:  &#8220;But, you want to come at me hard, I am going to come back harder at you. And then if I can’t jam my fingers down your throat, then I am not worth it. . . . But a little bit of respect.&#8221;  A couple years later, he followed up with some <a href="https://statmodeling.stat.columbia.edu/2022/01/29/fools-at-the-hoover-institution-the-gift-that-keeps-on-giving/">idiotic</a> statements about the covid vaccine.  Fine&#8211;the guy&#8217;s just a law professor, not a health economist or anyone else with relevant expertise here&#8211;the point is that Stanford appears to be stuck with him.  In Bhattacharya&#8217;s words, &#8220;Given this history, members of the public could be forgiven if they wonder whether any Stanford research can be trusted.&#8221;</p>
<p>This sort of guilt-by-Stanford-association would represent poor reasoning.  Just cos Stanford platforms idiots like Richard Epstein, it doesn&#8217;t mean we shouldn&#8217;t trust the research of serious scholars such as Rob Tibshirani and Jay Bhattacharya.  But I guess that members of the public could be forgiven if they show less trust in the Stanford brand.  Just as my Columbia affiliation is tarnished by my employer&#8217;s association with Mehmet Oz, Robert Hadden, and its willingness to fake its U.S. News numbers.  And Harvard is tarnished by its endorsement of various well-publicized bits of pseudoscience.</p>
<p>Reputation goes both ways.  By publishing Epstein&#8217;s uniformed commentary, Stanford’s leadership undermined public and scientific confidence, and then Bhattacharya had to pay some of the price for this undermined confidence.  That&#8217;s too bad, also too bad that he ended up on <a href="https://statmodeling.stat.columbia.edu/2021/01/29/team-stanford/">the advisory board</a> of an organization that gave the following advice:  &#8220;Currently, there is no one for whom the benefit would outweigh the risk of these [covid] vaccines&#8211;even the most vulnerable, elderly nursing home patients.&#8221;</p>
<p>The challenge is that legitimate arguments about policy responses under uncertainty get tangled with ridiculous claims such as that covid would only kill 500 Americans, or ridiculous policies such as removing the basketball hoops in the local park.  On one side, you had people spreading what can only be called denial (for example, the <a href="https://statmodeling.stat.columbia.edu/2021/01/29/team-stanford/">claim</a> that the pandemic was over in the summer of 2020); on the other side were public health authorities <a href="https://statmodeling.stat.columbia.edu/2024/11/19/behavioural-insights-team-for-the-loss/">playing the fear card</a> and keeping everyone inside.</p>
<p>So I can see how Bhattacharya was frustrated by the response to his study.  But he&#8217;s missing the mark when he misrepresents what I wrote, and when, elsewhere in his post, he disparages Stephanie Lee for doing reporting on his study.  We&#8217;re doing our jobs&#8211;I&#8217;m assessing the strength of statistical evidence, Lee is tracking down leads in the story&#8211;just like Bhattacharya is doing his job by making policy recommendations.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/22/stanford-medical-school-professor-misrepresents-what-i-wrote-but-i-kind-of-understand-where-hes-coming-from/feed/</wfw:commentRss>
			<slash:comments>90</slash:comments>
		
		
			</item>
		<item>
		<title>The true meaning of the alzabo</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/21/alzabo/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/21/alzabo/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Sat, 21 Dec 2024 14:17:25 +0000</pubDate>
				<category><![CDATA[Art]]></category>
		<category><![CDATA[Literature]]></category>
		<category><![CDATA[Miscellaneous Statistics]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51500</guid>

					<description><![CDATA[OK, I guess this is obvious, but it hadn&#8217;t occurred to me until recently. Nonexistent books Here&#8217;s the background: The other day we went to the Grolier Book Club, which had an exhibition on lost, incomplete, and imaginary books. The &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/21/alzabo/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>OK, I guess this is obvious, but it hadn&#8217;t occurred to me until recently.</p>
<p><strong>Nonexistent books</strong></p>
<p>Here&#8217;s the background:  The other day we went to the Grolier Book Club, which had an exhibition on lost, incomplete, and imaginary books.  The fun gimmick was that they made mock-ups of physical books representing various books that no longer exist or never were.  Examples included a lost play of Shakespeare, a never-finished novel by Sylvia Plath, the complete version of Kubla Khan, a monograph by Sherlock Holmes, the legendary Necronomicon, The Theory and Practice of Oligarchical Collectivism, by Emmanuel Goldstein, the code of the Bene Gesserit, and that famous original work which is a word-for-word remake of Don Quixote.  I was pleased to see a copy of the &#8220;original&#8221; Garden of Forking Paths but disappointed not to see anything by Nathan Zuckerman, and really disappointed not to see some version of The Book of Gold, but I understand, as there are so many lost, incomplete, and imaginary books that any library of them would overflow beyond the space allotted to the exhibition.</p>
<p>When I was at the Grolier looking at these real objects representing unreal books&#8211;actually, the objects were just the outsides of the books, I&#8217;m pretty sure they weren&#8217;t filled with text on their pages&#8211;I was thinking that there must be a wikipedia entry with a long list of famous books that don&#8217;t exist.  <a href="https://www.bookandsword.com/2024/08/10/wikipedia-culture-journalistic-culture-and-academic-culture/">Wikipedia has its flaws</a> (see also <a href="https://statmodeling.stat.columbia.edu/2011/11/25/bayes-wikipedia-update/">here</a> and <a href="https://statmodeling.stat.columbia.edu/2022/02/02/when-it-comes-to-error-correction-wikipedia-is-the-the-worst-form-of-government-except-for-all-those-other-forms-that-have-been-tried-from-time-to-time/">here</a>), but . . . a list of fake books, this seems like it would be right in Wikipedia&#8217;s wheelhouse, kind of like if you want the plot of every Simpsons episode or a list of all the flavors of Pop Tarts.</p>
<p>Some googling led to the wikipedia page for <a href="https://en.wikipedia.org/wiki/Fictional_book">&#8220;fictional book&#8221;</a>.  It&#8217;s not as comprehensive as I was expecting, but it does include The Grasshopper Lies Heavy&#8212;I&#8217;d forgotten about that one!  Wikipedia also has <a href="https://en.wikipedia.org/wiki/Category:Fictional_books">this list</a>, which is even more disappointing.</p>
<p>What&#8217;s up, Wikipedia?  I was gonna say that the editors are so laser-focused on modern pop culture that they can&#8217;t be bothered to collect a list of old-fashioned <em>books</em> . . . but then I checked out the page on <a href="https://en.wikipedia.org/wiki/Lost_literary_work">&#8220;lost literary works&#8221;</a>, which offers an impressively comprehensive list starting with antiquity and going through the centuries since then, ending with this:</p>
<blockquote><p>Terry Pratchett&#8217;s unfinished works were destroyed in 2017 after his death, fulfilling his last will; his computer hard drive containing his unfinished works was deliberately crushed by a steamroller.</p></blockquote>
<p>The list of lost works includes novels by L. Frank Baum, plays by James Joyce, August Strindberg, and Leon Trotsky (!), treatises by Adam Smith, all sorts of things.  It&#8217;s a list worth reading.  Wikipedia also has an entry on <a href="https://en.wikipedia.org/wiki/Unfinished_creative_work">unfinished creative work</a>, which includes things like Mozart&#8217;s requiem and Schubert&#8217;s eighth symphony as well as various books.</p>
<p><strong>The alzabo</strong></p>
<p>As noted above, when I saw nonexistent books, I thought of The Book of Gold, and that made me think of the alzabo, that fictional animal who, if it eats a person, will ingest his or her personality as well, and then if you kill the alzabo, prepare it in a certain way, and swallow its extract, you will retain the memories of the person who&#8217;d been eaten.</p>
<p>Seeing the exhibition of the books gave me the sudden understanding that the alzabo is, very directly, a metaphor for literature:</p>
<p>&#8211; When, as an author, you pour your memories, ideas, and emotions into a book, that is like feeding yourself to the alzabo.  As Paul Gallico <a href="https://quoteinvestigator.com/2011/09/14/writing-bleed/">said</a>, &#8220;It is only when you open your veins and bleed onto the page a little that you establish contact with your reader.&#8221;</p>
<p>&#8211; When, as a reader, you dive into a book, this is like eating the alzabo.  The characters, themes, and events of the book become part of your memory.</p>
<p>In the Book of the New Sun, there are times when Severian is overwhelmed by the memories he&#8217;s absorbed from Thecla, and there&#8217;s a point where he gets lost among the accumulated memories of all the past autarchs.  When I read that, I recall thinking this was implausible (even in the science-fiction context of a willing suspension of disbelief)&#8211;how would there be room in his head for the memories of thousands of people?&#8211;but then I realized that my head contains memories of many thousands of books that I&#8217;ve read (not to mention TV shows and various other story-delivering mechanisms), so, yeah, that&#8217;s how it works.  The alzabo is a beautifully poetic representation of this two-stage transfer, first from author to book, then from book to reader.</p>
<p>In one of history&#8217;s greatest fish-out-of-water stories, a retired Hollywood actor became president of the United States.  Ronald Reagan was notorious for confusing his memories of movies with reality&#8211;but, at some level, all of us can relate to this, as we all &#8220;know&#8221; fictional characters who seem realer than people we know, we remember fictional stories that are scarier and more tearjerking than real events in our lives, and so forth.</p>
<p><strong>Connection to statistics</strong></p>
<p>And then, writing this, I realize that something similar happens in statistics!  Measurements are performed that abstract complex reality into &#8220;data,&#8221; which are then analyzed by the researcher, who uses these data to reconstruct the world.  Wow.  I&#8217;d never thought of it that way before.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/21/alzabo/feed/</wfw:commentRss>
			<slash:comments>12</slash:comments>
		
		
			</item>
		<item>
		<title>Delicate language for talking about statistical guarantees</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/20/delicate-language-for-talking-about-statistical-guarantees/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/20/delicate-language-for-talking-about-statistical-guarantees/#comments</comments>
		
		<dc:creator><![CDATA[Jessica Hullman]]></dc:creator>
		<pubDate>Fri, 20 Dec 2024 17:15:21 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Miscellaneous Statistics]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51489</guid>

					<description><![CDATA[This is Jessica. As I’ve been reading papers on uncertainty quantification in machine learning (related to topics like conformal prediction or calibration), I’ve been reflecting on the language choices that authors make.  It started when I was reading a paper &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/20/delicate-language-for-talking-about-statistical-guarantees/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><span style="font-weight: 400">This is Jessica. As I’ve been reading papers on uncertainty quantification in machine learning (related to topics like conformal prediction or calibration), I’ve been reflecting on the language choices that authors make. </span></p>
<p><span style="font-weight: 400">It started when I was reading a paper about an approach that uses set-aside data labeled with human reports to learn a threshold on a function that predicts the &#8220;alignment&#8221; (e.g., similarity) of predictions with those reports. They do this to ensure that new predictions are sufficiently aligned. There are many new methods for quantifying prediction uncertainty or controlling prediction risk in ML that have a similar flavor &#8212; they involve learning either thresholds (e.g., conformal prediction) or adjustments (e.g., posthoc calibration algorithms) on held-out calibration data that are then applied to predictions. Some of them come with &#8220;statistical guarantees&#8221;: If the data used for calibration is a good approximation of future data (e.g., is i.i.d. or at least exchangable), then we can expect calibration in the future.</span></p>
<p>Anyway, I paused when I got to this sentence:</p>
<blockquote><p><span style="font-weight: 400">new reports are selected if their predicted alignment scores exceed a data-driven threshold, which is delicately set with Dcal such that the FDR is strictly controlled at the desired level</span></p></blockquote>
<p><span style="font-weight: 400"> because I found the choice of the phrase “delicately set” interesting. It seems like a good way to describe this kind of procedure: naturally there are assumptions on which the expectation about controlling the FDR rate depend, like statistical exchangeability between the calibration data and new instances, that might not hold. By using a term like </span><span style="font-weight: 400">“delicately” the authors signal some fragility. </span></p>
<p><span style="font-weight: 400">It makes me think of all of the other words that could be used to signal tentativeness or epistemic uncertainty when discussing such methods: “tenuously,” “gingerly,” “warily,“ or, if you want to sound like you just studied for the GRE, “frangibly.” Maybe a bit of poetic license would not be a bad thing when describing our expectations about such methods. It would certainly make for a more entertaining reading experience.</span></p>
<p><span style="font-weight: 400">However, the fact that the above statement continues </span><span style="font-weight: 400">“&#8230; </span><span style="font-weight: 400">such that the FDR is strictly controlled at the desired level”</span><span style="font-weight: 400"> makes it sound a lot less fallible. This phrasing is more typical of the unconditional way in which such methods are often described.  </span></p>
<p><span style="font-weight: 400">Here’s another example, where the guarantee is about human behavior: </span></p>
<blockquote><p><span style="font-weight: 400">we develop an efficient and near-optimal search method to find the conformal predictor under which the expert is guaranteed to achieve the greatest accuracy with high probability.</span></p></blockquote>
<p><span style="font-weight: 400">But can we ever really guarantee that future human behavior will remain consistent with the past? </span></p>
<p><span style="font-weight: 400">As I was writing this post, I noticed that Andrew has previously <a href="http://The more I read these kinds of papers, the more I agree with Andrew that using the term &quot;guarantees&quot; is not helping.">blogged on his dislike of the term &#8220;guarantees&#8221; </a>because it hide assumptions. </span><span style="font-weight: 400">I agree that it&#8217;s problematic to talk about guarantees unconditionally, as if they cannot be violated even when the methods are applied in practice. This is why the unqualified <a href="https://statmodeling.stat.columbia.edu/2024/11/01/calibration-is-sometimes-sufficient-for-trusting-predictions-what-does-this-tell-us-when-human-experts-use-model-predictions/">statements about calibration being sufficient for good decisions bug me</a> so much. </span><span style="font-weight: 400">Andrew&#8217;s post concludes by saying that it&#8217;s not so bad to talk about guarantees if you&#8217;re writing to a statistics and ML audience, since they may expect that. </span></p>
<p><span style="font-weight: 400">But the more time I spend reading papers involving uncertainty or calibration guarantees, the more I find myself wishing there was a little more care taken in how their statistical properties are discussed, particularly when talking about practically-oriented methods that assume future data is like past data. Having poked around recent papers in domains outside of CS on topics like calibration and conformal prediction, it seems that the <a href="https://statmodeling.stat.columbia.edu/2024/06/14/loving-hating-and-sometimes-misinterpreting-conformal-prediction-for-medical-decisions/">nuance is not necessarily obvious</a> when people go to apply the methods in practice. So t</span><span style="font-weight: 400">hese days I&#8217;m trying to be more careful with my own word choices. For example, beyond trying to avoid flagrant use of the word guarantee, add the word “expected” more often. E.g., “under which the expert is expected to achieve the greatest accuracy…” or “delicately selected so that the FDR is expected to be strictly controlled at the desired level.” It’s minor but I think it better emphasizes that our assumptions may not always don&#8217;t hold, because they were unrealistic and/or because in practice we’re dealing with singular events rather than hypothetical replications. </span></p>
<p><span style="font-weight: 400">Another idea would be to deliberately choose weaker synonyms for guarantees, like “this method promises that the expected FDR will be less than&#8230;” Sounds a little odd. But maybe it’s a good thing for descriptions of methods to take the reader back to childhood when their so-called friend “promised” not to tell on them or their parent “promised” not to snoop in their room! </span><span style="font-weight: 400">Or as Paul Alper <a href="https://statmodeling.stat.columbia.edu/2024/07/05/you-can-guarantee-that-the-term-statistical-guarantee-will-irritate-me-heres-why-and-lets-go-into-some-details/#comment-2375093">suggests</a>, say &#8220;statistical oomph&#8221; instead of guarantees, which I think captures the same vibe of excitement and clout but sounds so much sillier.</span></p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/20/delicate-language-for-talking-about-statistical-guarantees/feed/</wfw:commentRss>
			<slash:comments>8</slash:comments>
		
		
			</item>
		<item>
		<title>&#8220;Accounting for Nonresponse in Election Polls: Total Margin of Error&#8221;</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/20/accounting-for-nonresponse-in-election-polls-total-margin-of-error/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/20/accounting-for-nonresponse-in-election-polls-total-margin-of-error/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Fri, 20 Dec 2024 14:13:33 +0000</pubDate>
				<category><![CDATA[Bayesian Statistics]]></category>
		<category><![CDATA[Miscellaneous Statistics]]></category>
		<category><![CDATA[Multilevel Modeling]]></category>
		<category><![CDATA[Political Science]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=50948</guid>

					<description><![CDATA[Jeff Dominitz and Chuck Manski write: The potential impact of nonresponse on election polls is well known and frequently acknowledged. Yet measurement and reporting of polling error has focused solely on sampling error, represented by the margin of error of &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/20/accounting-for-nonresponse-in-election-polls-total-margin-of-error/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>Jeff Dominitz and Chuck Manski <a href="https://arxiv.org/abs/2407.19339">write</a>:</p>
<blockquote><p>The potential impact of nonresponse on election polls is well known and frequently acknowledged. Yet measurement and reporting of polling error has focused solely on sampling error, represented by the margin of error of a poll. Survey statisticians have long recommended measurement of the total survey error of a sample estimate by its mean square error (MSE), which jointly measures sampling and non-sampling errors. Extending the conventional language of polling, we think it reasonable to use the square root of maximum MSE to measure the total margin of error. This paper demonstrates how to measure the potential impact of nonresponse using the concept of the total margin of error, which we argue should be a standard feature in the reporting of election poll results. We first show how to jointly measure statistical imprecision and response bias when a pollster lacks any knowledge of the candidate preferences of non-responders. We then extend the analysis to settings where the pollster has partial knowledge that bounds the preferences of non-responders.</p></blockquote>
<p>Good stuff.  This relates to two of my papers on nonsampling errors and differential nonresponse:</p>
<p><a href="http://stat.columbia.edu/~gelman/research/published/polling-errors.pdf">[2018] Disentangling bias and variance in election polls. Journal of the American Statistical Association 113, 607-614.</a> (Houshmand Shirani-Mehr, David Rothschild, Sharad Goel, and Andrew Gelman)</p>
<p><a href="http://stat.columbia.edu/~gelman/research/published/nonresp2.pdf">[1998] Modeling differential nonresponse in sample surveys. Sankhya 60, 101-126.</a> (Thomas C. Little and Andrew Gelman)</p>
<p>It&#8217;s funny about that last paper because Tom Little and I did that project nearly 30 years ago and I forgot about it even while doing <a href="http://stat.columbia.edu/~gelman/research/published/swingers.pdf">applied research</a> on differential nonresponse.  I think there&#8217;s more to be done here, integrating the different perspectives here.  I like the idea of modeling the nonsampling error rather than just treating it as another error term.</p>
<p>On the general point of reporting surveys and margins of errors, I recommend <a href="https://www.tandfonline.com/doi/abs/10.1080/09332480.1993.11882474">Ansolabehere and Belin&#8217;s paper</a> from 1993.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/20/accounting-for-nonresponse-in-election-polls-total-margin-of-error/feed/</wfw:commentRss>
			<slash:comments>28</slash:comments>
		
		
			</item>
		<item>
		<title>How did the press do on that &#8220;black spatula&#8221; story?  Not so great.</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/19/how-did-the-press-do-on-that-black-spatula-story/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/19/how-did-the-press-do-on-that-black-spatula-story/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Thu, 19 Dec 2024 15:25:27 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=51486</guid>

					<description><![CDATA[Act 1: The journal Chemosphere publishes a research article finding that black plastic kitchen items contain dangerous toxins arising from reprocessing of plastics that contained carcinogenic flame retardants. This result is publicized by the advocacy group that helped conduct the &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/19/how-did-the-press-do-on-that-black-spatula-story/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p><strong>Act 1:</strong>  The journal Chemosphere publishes a research article finding that black plastic kitchen items contain dangerous toxins arising from reprocessing of plastics that contained carcinogenic flame retardants.  This result is publicized by the advocacy group that helped conduct the study, and it is picked up in the major media, with headlines such as, &#8220;Black-colored plastic used for kitchen utensils and toys linked to banned toxic flame retardants,&#8221; as <a href="https://www.cnn.com/2024/10/01/health/flame-retardant-black-plastic-wellness/index.html">CNN reported</a> from 1 Oct.  Various news organizations pick up the story and give recommendations about not reusing plastic containers, etc.</p>
<p><strong>Act 2:</strong>  It turns out that the journal article had a factor-of-10 error, where a given exposure was stated to be 80% of the legal limit for a certain toxin, but it was really only 8%.</p>
<p>We <a href="https://statmodeling.stat.columbia.edu/2024/12/13/how-a-simple-math-error-sparked-a-panic-about-black-plastic-kitchen-utensils/">wrote about this the other day</a>, citing, linking, and quoting from <a href="https://nationalpost.com/news/canada/black-plastic">an article of 11 Dec</a> from Canada&#8217;s National Post newspaper, that reported:</p>
<blockquote><p>Plastics rarely make news like this . . . the media uptake was enthusiastic on a paper published in October in the peer-reviewed journal Chemosphere.</p>
<p>“Your cool black kitchenware could be slowly poisoning you, study says. Here’s what to do,” said the LA Times. “Yes, throw out your black spatula,” said the San Francisco Chronicle. Salon was most blunt: “Your favorite spatula could kill you,” it said.</p>
<p>The study, by researchers at the advocacy group Toxic-Free Future . . . estimated that using contaminated kitchenware could cause a median intake of 34,700 nanograms per day of Decabromodiphenyl ether, known as BDE-209. . . . The trouble is that, in the study’s section on “Health and Exposure Concerns,” the researchers said this number, 34,700, “would approach” the reference dose given by the United States Environmental Protection Agency. . . .</p>
<p>The paper correctly gives the reference dose for BDE-209 as 7,000 nanograms per kilogram of body weight per day, but calculates this into a limit for a 60-kilogram adult of 42,000 nanograms per day. . . . But 60 times 7,000 is not 42,000. It is 420,000. This is what [McGill University’s] Joe Schwarcz noticed. . . .</p></blockquote>
<p>The article goes on to interview Schwarcz about the spatula claim and general issues of summarizing uncertainty.</p>
<p>The starting point of Act 2 of our story seems to have been <a href="https://www.mcgill.ca/oss/article/critical-thinking-health-and-nutrition/are-black-plastic-spatulas-and-serving-spoons-safe-use">this press release</a> by Schwarcz from 6 Dec on the McGill University website, entitled, &#8220;Are Black Plastic Spatulas and Serving Spoons Safe to Use? For me, the risk would not be enough to discard them, but science should find a way to keep flame retardants out of such items.&#8221;</p>
<p>It&#8217;s a complicated problem.  Indeed, only just now I&#8217;m realizing that &#8220;median intake&#8221; is a strange thing to look at when summarizing a public health risk.</p>
<p><strong>Act 3:</strong>  The error was corrected in the journal and in a press release by the advocacy organization, but in a minimizing way:  &#8220;it is important to note that this does not impact our results.&#8221;  This was mentioned in the above-linked National Post article.</p>
<p><strong>Act 4:</strong>  Several news organizations followed up on the correction.</p>
<p>And that&#8217;s the subject of today&#8217;s post.  How did the news organizations do?</p>
<p>I&#8217;d say their record is mixed.<br />
<span id="more-51486"></span><br />
I&#8217;ll go through them in the order they came up in my Google searches: </p>
<p><a href="https://www.sfgate.com/food/article/throw-away-black-spatulas-sf-19986362.php">Here&#8217;s SF Gate</a>, a website that was <a href="https://www.sfgate.com/home/article/about-sfgate-15613713.php">formerly part of</a> the San Francisco Chronicle: </p>
<blockquote><p>Did you read that black spatulas were poisoning you and everyone you love and then throw away multiple kitchen utensils in a blind panic? . . . Publications ranging from the Atlantic (“Throw Out Your Black Plastic Spatula”) to the Los Angeles Times (“Your cool black kitchenware could be slowly poisoning you, study says”) to Salon.com (“Your favorite spatula could kill you”) all spread the panic. . . . But then Joe Schwarcz, director of McGill University’s Office for Science and Society, noticed a mistake in the study. . . .</p></blockquote>
<p>The SF Gate article, which is dated 17 Dec, has some strong similarities to the National Post article&#8211;I&#8217;m not saying it&#8217;s plagiarized, exactly, more that they seem to have read that earlier news article and rewritten it.  The good news is they kept the credit to Schwarcz; the bad news is . . . well, let&#8217;s see if you notice.  Read carefully:</p>
<p>From the National Post:  &#8220;&#8216;Your cool black kitchenware could be slowly poisoning you, study says. Here’s what to do,&#8217; said the LA Times. &#8216;Yes, throw out your black spatula,&#8217; said the San Francisco Chronicle. Salon was most blunt: &#8216;Your favorite spatula could kill you,&#8217; it said.&#8221;</p>
<p>From SF Gate:  &#8220;Publications ranging from the Atlantic (&#8216;Throw Out Your Black Plastic Spatula&#8217;) to the Los Angeles Times (&#8216;Your cool black kitchenware could be slowly poisoning you, study says&#8217;) to Salon.com (&#8216;Your favorite spatula could kill you&#8217;) all spread the panic.&#8221;</p>
<p>SF Gate kept the LA Times and Salon.com, they added the Atlantic, but they removed the San Francisco Chronicle.  Kind of funny to get rid of the only local angle, no?  Especially given that the SF Gate article has the title, &#8220;Turns out everyone in SF didn&#8217;t need to throw away their black spatulas.&#8221;</p>
<p>I can&#8217;t figure out why SF Gate didn&#8217;t mention the San Francisco Chronicle.  Here are two, somewhat contradictory, guesses:</p>
<p>1.  SF Gate and the Chronicle are competitors and so, wherever possible, they will avoid linking to or mentioning each other.</p>
<p>2.  SF Gate and the Chronicle are sister organizations and so SF Gate removed any mention of the Chronicle in their article so as not to cause any embarrassment to that newspaper.</p>
<p>I don&#8217;t know which it is; either way, sure, it&#8217;s not the world&#8217;s biggest scandal but it&#8217;s kind of uncool.</p>
<p>In any case, SF Gate did not credit the National Post at all, which seems really bad given that it looks like they were just rewriting that National Post story from a week earlier.  Again, I&#8217;m not saying it was plagiarism; it&#8217;s just bad practice to not credit an earlier source.</p>
<p>Google also <a href="https://web.archive.org/web/20241217125042/https://www.sfchronicle.com/health/article/black-plastic-study-problem-19984247.php">turns up an article</a>, &#8220;Viral study about black plastic spatulas had a big math problem,&#8221; from the San Francisco Chronicle itself.  This one also cites Schwarcz, but not the National Post, even though, again, it seems to be a recycling (ha ha, geddit?) of that earlier news article.</p>
<p>On the plus side, the Chronicle does explicitly say that they had covered the earlier story, and they link to that story, although without reproducing their incendiary (ha ha, geddit?) headline, &#8220;Yes, throw out your black spatula.&#8221;</p>
<p>Here&#8217;s <a href="https://www.nytimes.com/2024/12/10/well/black-plastic-health.html">the New York Times</a>, my hometown newspaper, with an article entitled, &#8220;Do I Really Need to Throw Out My Black Plastic Spatula? A new study detected dangerous chemicals in a variety of household items. But experts say the health risks aren’t clear-cut.&#8221;  In true Gray Lady fashion, they&#8217;ve worked to make their headline and sub-headline as boring as possible&#8212;and they&#8217;ve pretty much succeeded!</p>
<p>This Times article does the worst job of reporting we&#8217;ve seen so far!  In some ways it&#8217;s good:  it takes a measured tone and includes interviews with multiple experts.  But . . . the article was published 10 Dec and updated 17 Dec, and it doesn&#8217;t mention the factor-of-10 problem at all!  Even though Schwarcz&#8217;s press release appeared on 6 Dec.  OK, fair enough&#8211;the reporter didn&#8217;t go to the trouble of a careful internet search that could&#8217;ve found the press release, and as of 10 Dec, when the NYT article was first published, the National Post article hadn&#8217;t yet appeared.  It was really the National Post that broke the story.  But it says right there on the Times article that it was updated on 17 Dec, indeed that it appeared in the physical newspaper on that day.  By then, even a cursory internet search on some variant of *black plastic study* would&#8217;ve revealed the problem.</p>
<p>Back on 14 Nov, the Times had run <a href="https://www.nytimes.com/wirecutter/reviews/toxic-black-plastic-kitchen-alternatives/">an article</a>, &#8220;Black Plastic Kitchen Tools Might Expose You to Toxic Chemicals. Here’s What to Use Instead.&#8221;  They didn&#8217;t know about the factor-of-10 error, and they pretty much buy the &#8220;Toxic-Free Future&#8221; alarmist pitch straight-up, no skepticism.  On the plus side, unlike all the other sources, this Times article has some useful suggestions, as they discuss metal, silicone, and wood alternatives.</p>
<p>On the positive side, <a href="https://slate.com/technology/2024/11/black-plastic-spatula-chemicals-flame-retardants.html">here&#8217;s Slate</a>, which did very well on this story. Back on 4 Nov, long before any of the scandal had occurred, they ran a story, &#8220;I’m Not Throwing Away My Black Plastic Spatula. Yes, they can have scary chemicals in them. But let’s take a closer look at the research,&#8221; with lots of details going into the science.  And, as a bonus, on 16 Dec they added an update at the very beginning of their story, &#8220;Black plastic is even less worrisome than it initially seemed—one of the papers sparking concern <a href="https://nationalpost.com/news/canada/black-plastic">contained a math error</a>,&#8221; with that last link going back to . . . the National Post story from 11 Dec!</p>
<p>Good job, Slate!  Excellent science reporting.  And I say this not just as an <a href="http://stat.columbia.edu/~gelman/media/">occasional</a> science reporter for Slate.</p>
<p>The Los Angeles Times, to its credit, also got on the bandwagon, with a story dated 19 Dec entitled &#8220;Your black plastic kitchen utensils aren’t so toxic after all. But you should still toss them, group says,&#8221; and continuing:</p>
<blockquote><p>Published in the peer-reviewed journal Chemosphere, experts from the nonprofit Toxic-Free Future said they detected flame retardants and other toxic chemicals in 85% of 203 items made of black plastic including kitchen utensils, take-out containers, children’s toys and hair accessories. . . .</p>
<p>But in an update to the study, the authors say they made an error in their calculations and the real levels were “an order of magnitude lower” than the EPA’s thresholds. The error was discovered by Joe Schwarcz, director of McGill University’s Office for Science and Society in Canada.</p>
<p>In a blog post, Schwarcz explained that the Toxin-Free Future scientists miscalculated the lower end of what the EPA considered a health risk through a multiplication error. Instead of humans being potentially exposed to a dose of toxic chemicals in black plastic utensils near the minimum level that the EPA deems a health risk, it’s actually about one-tenth of that.</p></blockquote>
<p>The L.A. Times didn&#8217;t link to or credit the National Post, but they did credit Schwarcz, which is important.  Oddly enough, though, they did not link to Schwarcz&#8217;s press release (which they label as a blog, which is I don&#8217;t think is quite accurate, but maybe it&#8217;s close enough).  The L.A. Times does link to their earlier article, but they do a bad thing by (a) not pointing out that it was their newspaper that fell for that earlier hype (to know it was the L.A. Times, you&#8217;d need to hover over or click on that link) and (b) not pointing out their earlier, alarmist, headline, &#8220;Your cool black kitchenware could be slowly poisoning you, study says. Here’s what to do.&#8221;</p>
<p>Some further internet searching took me to the world of news aggregators&#8211;those sites that manage to show up in your google search but don&#8217;t do any reporting themselves.  The site msn.com offered two articles dated today, <a href="https://www.msn.com/en-za/news/other/potential-cancer-causing-chemicals-have-been-found-in-black-spatulas-and-takeaway-containers-is-it-time-to-throw-them-out/ar-AA1w9Ixy">one from</a> the South African site DW and <a href="https://www.msn.com/en-us/health/medical/maybe-it-s-not-time-to-chuck-black-spatulas/ar-AA1vZdLq">one from</a> the U.S.? site Newser.  The latter article is just a rewrite of the National Post article.  To its credit, it cites and links to the National Post, so in that sense, it&#8217;s the best of the bunch.  On the other hand, the whole thing is kind of sketchy:  Newser describes itself as &#8220;a human + ai experiment,&#8221; and what it&#8217;s doing is paraphrasing an existing article and then putting it on some new platform with a bunch of ads, which is then scraped by msn.com, which adds no value but is then itself included in a google search.  Lots of middlemen in this system, and at no stage is any value added.</p>
<p><a href="https://www.scientificamerican.com/article/study-miscalculation-has-everyone-talking-about-black-plastic-spatulas-again/">Here&#8217;s Scientific American</a>, with a story dated 18 Dec, &#8220;Any Level of Flame Retardants in Black Plastic Spatulas Concerns Scientists. The scientists behind a popular study on the health effects of flame retardants in black plastic cooking utensils and toys made a calculation error but still say their revised findings are alarming.&#8221;</p>
<p>I&#8217;m not so happy about this story, for two reasons.  First, yes, they mention the factor-of-10 error but without citing or linking to Schwarcz at all, indeed not mentioning him at all.  Second, they give lots of quotes from the advocacy group, pretty much just following their spin.  I guess this fits into Scientific American&#8217;s recent reputation as being a politicized anti-industry propaganda sheet.</p>
<p>From the other side, <a href="https://www.plasticstoday.com/medical/the-case-of-the-black-plastic-spatula">here&#8217;s Plastics Today</a>, representing the plastics industry, or some segment of it.  Their article is called, &#8220;The Case of the Black Plastic Spatula. Did you throw yours out? That might have been premature.&#8221;  They don&#8217;t mention Joe Schwarcz, but they correctly do cite and link to the National Post.</p>
<p><a href="https://www.usatoday.com/story/news/nation/2024/12/18/black-plastic-kitchen-utensils-study-risk/77049209007/">USA Today</a> reports the correction, which is fine, but they only cite the advocacy group that published the paper, without crediting Schwarcz.  They link to the advocacy group&#8217;s updated <a href="https://toxicfreefuture.org/update-study-correction-does-not-impact-conclusion/">press release</a>, which minimizes their factor-of-10 mistake (&#8220;The error does not impact the study’s findings, recommendations, or conclusions,&#8221; which makes you wonder why they bothered putting the number in their paper in the first place, also makes you wonder if they&#8217;d said it didn&#8217;t impact anything had the error been a factor of 10 in a direction that did not favor their advocacy) and, yes, they say &#8220;We are grateful to the scientists who brought this forward and we apologize for the error,&#8221; but, no, they do not link to Schwarcz&#8217;s report or credit him in any way.</p>
<p>As you might expect, the tech news site Gizmodo <a href="https://gizmodo.com/embarrassing-math-error-scared-the-pants-off-people-over-toxic-black-kitchenware-2000540158">does a better job</a>, citing Schwarcz and linking to the National Post article.</p>
<p>Bill Bottenberg pointed me to an article dated 16 Dec in the news site Ars Technica, &#8220;Huge math error corrected in black plastic study; authors say it doesn’t matter. Correction issued for black plastic study that had people tossing spatulas.&#8221;  It&#8217;s pretty bad:  zero credit to Schwarcz or the National Post, it just goes over the original published paper and the correction notice from the journal.</p>
<p>Regarding the substance of the matter, Bottenberg writes:</p>
<blockquote><p>Being a natural born physical chemist I noticed the earlier claim about the dangers of black plastic in the kitchen. Of course I examined immediately my kitchen tool drawers and found: narrow spatula with a lotta holes, wide spatula with a flat front edge &#8211; good for flipping omeletes, nice slotted spoon good for removing stuff from a boiling broth for various reasons, and a deep wide spoon good for serving from a soup or stew. These were all as black and plasticky as could be. I looked at them and let them be.</p>
<p>Why would I do this? Well, maybe a feeling that the mass transfer processes in water (with stuff) didn&#8217;t seem likely at the temperature ranges that might be encountered. But then, the mass transfer rates in dang hot oil (with stuff) might be meaningful. These considerations might be for aromatic molecules, maybe for nasty bromine containing organics aromatic or not.</p>
<p>What a puzzle. I closed the drawers and continued using them as I mulled over the possibilities.</p>
<p>This is how we human beings do stuff. Either we throw them all out, keep them because we don&#8217;t care, or just think for a while about the ramifications&#8230;</p>
<p>To the current points you have been making, which I really enjoy and appreciate &#8211; the authors have admitted to the mistake, but not so much as to change their recommendations. It&#8217;s a strange world.</p>
<p>Ha ha, the jokes on me. Now I have to go back and really read the original paper, rethink the Ars Technica article and figure out what&#8217;s next.</p>
<p>I&#8217;m not emptying the drawers of black plastic&#8230;</p></blockquote>
<p>My recommendations to Bottenberg are:</p>
<p>1.  Don&#8217;t &#8220;rethink the Ars Technica article.&#8221;  Just go back to the original sources.  And I&#8217;d recommend spending more time on Gizmodo and less time on Ars Technica (and zero time on Scientific American!).</p>
<p>2.  Use a metal ladle to serve your soup, wooden spoons for stirring, a metal spatula for frying on your usual pans, and a silicone spatula for your nonstick pans.  If you don&#8217;t have any of these items, then, sure, no rush to get them all at once, but you must already have some of these in your kitchen already, no?</p>
<p><strong>Summary</strong></p>
<p>The news organizations did a reasonable job covering the correction, except for four things:</p>
<p>1.  Many fewer outlets reported on the error than on the original story.</p>
<p>2.  Even when reporting the error, they mostly relied too strongly on the spin provided by the advocacy organization.</p>
<p>3.  They mostly did not cite the National Post, even though that was the newspaper that broke the story.  Unfortunately, this seems to be standard practice, most notoriously with the New York Times but with other newspapers as well, to not credit the reports from other newspapers that got to a story first.  <a href="https://statmodeling.stat.columbia.edu/2023/08/12/nyt-does-some-rewriting-without-attribution-i-guess-this-is-standard-in-journalism-but-it-seems-unethical-to-me/">Here&#8217;s a particularly bad example</a> from last year of the NYT retelling a story without acknowledging the original source, which was a Canadian newspaper.</p>
<p>4.  Many of the outlets did not credit Joe Schwarcz, the person who discovered the error.  This was a consequence of items 2 and 3 above, as Schwarcz was not credited in the correction notice and press release from the authors of the study.</p>
<p>This is not plagiarism, but a similar issue arises here, something that <a href="http://stat.columbia.edu/~gelman/research/published/GelmanBasbollAmericanScientist.pdf">Thomas Basbøll and I have written about</a>: When a news article obscures is sources, information is lost, and the reader pays the price.</p>
<p>In this case, it&#8217;s clear that several of the followup articles were directly based on the National Post article; by not mentioning or linking to that earlier reporting, they were covering their tracks and making it harder for readers to see beyond the spin being provided by the advocacy organization.</p>
<p>I&#8217;m not saying here that plastic kitchen items are safe (or that they&#8217;re not); what I&#8217;m talking about is the reporting.</p>
<p>Also, don&#8217;t trust Scientific American.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/19/how-did-the-press-do-on-that-black-spatula-story/feed/</wfw:commentRss>
			<slash:comments>18</slash:comments>
		
		
			</item>
		<item>
		<title>He took public funds and falsified his data.  Are they gonna make him pay back the $19 million?</title>
		<link>https://statmodeling.stat.columbia.edu/2024/12/18/are-they-gonna-make-him-pay-back-the-19-million/</link>
					<comments>https://statmodeling.stat.columbia.edu/2024/12/18/are-they-gonna-make-him-pay-back-the-19-million/#comments</comments>
		
		<dc:creator><![CDATA[Andrew]]></dc:creator>
		<pubDate>Wed, 18 Dec 2024 14:33:31 +0000</pubDate>
				<category><![CDATA[Miscellaneous Science]]></category>
		<category><![CDATA[Political Science]]></category>
		<category><![CDATA[Zombies]]></category>
		<guid isPermaLink="false">https://statmodeling.stat.columbia.edu/?p=50928</guid>

					<description><![CDATA[As a former University of Maryland student, I&#8217;m incensed by this story. From Retraction Watch, this story about a professor of Biochemistry and Molecular Biology: The ORI finding stated Eckert “engaged in research misconduct in research supported by” every NIH &#8230; <a href="https://statmodeling.stat.columbia.edu/2024/12/18/are-they-gonna-make-him-pay-back-the-19-million/">Continue reading <span class="meta-nav">&#8594;</span></a>]]></description>
										<content:encoded><![CDATA[<p>As a former University of Maryland student, I&#8217;m incensed by <a href="https://retractionwatch.com/2024/08/13/former-maryland-dept-chair-with-19-million-in-grants-faked-data-in-13-papers-feds-say/">this story</a>.  From Retraction Watch, this story about a professor of Biochemistry and Molecular Biology:</p>
<blockquote><p>The ORI finding stated Eckert “engaged in research misconduct in research supported by” every NIH grant on which he served as principal investigator, totaling more than $19 million. The finding also lists multiple “Center Core Grants” worth hundreds of millions for shared resources and facilities at research centers. . . . </p>
<p>According to ORI’s findings, Eckert erased a band in one of the paper’s figures “to falsely show a favorable result.” </p>
<p>In the 13 papers and two grant applications, Eckert used and reused images “representing unrelated experiments, with or without manipulating them, and falsely relabeling them as data representing different proteins and/or experimental results,” ORI found.</p></blockquote>
<p>I don&#8217;t really know what to think about the Center Core Grants, but the 19 million dollars he should have to pay back!  OK, maybe he doesn&#8217;t have $19 million cash on hand, but there&#8217;s gotta be a system for this, right?  Maybe they start by repoing his car, freezing his bank account, and selling his house, then he&#8217;s allowed some reasonable amount to live on . . . hmmm, it appears that the NIH predoc salary level is 28,224 per year, so they could let him keep that much . . . then he could try to come up with a payment plan.  I dunno.  They throw drug dealers in prison, but that would just cost the taxpayer more money.  Better to just put an ankle monitor on him and take away his internet access. . . .</p>
<p>Ummm, let&#8217;s read on:</p>
<blockquote><p>Eckert agreed to forgo contracting with the federal government or receiving government funding for eight years . . . Eckert also agreed not to serve on any advisory or peer review committees for the U.S. Public Health Service, which includes the NIH, for eight years.</p></blockquote>
<p>OK, maybe after eight years if he shows some evidence of rehabilitation, maybe.  But what about the $19 million?  Paying that back should be the absolute minimum requirement, no?</p>
]]></content:encoded>
					
					<wfw:commentRss>https://statmodeling.stat.columbia.edu/2024/12/18/are-they-gonna-make-him-pay-back-the-19-million/feed/</wfw:commentRss>
			<slash:comments>48</slash:comments>
		
		
			</item>
	</channel>
</rss>
