<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Study Hacks - Decoding Patterns of Success - Cal Newport</title>
	<atom:link href="https://calnewport.com/blog/feed/" rel="self" type="application/rss+xml" />
	<link>https://calnewport.com/blog/</link>
	<description>Computer Scientist &#38; Bestselling Author</description>
	<lastBuildDate>Mon, 27 Apr 2026 02:24:25 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	

<image>
	<url>https://calnewport.com/wp-content/uploads/2022/10/cropped-cal-newport-favicon-512x512-1-32x32.png</url>
	<title>Study Hacks - Decoding Patterns of Success - Cal Newport</title>
	<link>https://calnewport.com/blog/</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>Who Asked For This?</title>
		<link>https://calnewport.com/who-asked-for-this/</link>
					<comments>https://calnewport.com/who-asked-for-this/#comments</comments>
		
		<dc:creator><![CDATA[Study Hacks]]></dc:creator>
		<pubDate>Mon, 27 Apr 2026 10:00:00 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<guid isPermaLink="false">https://calnewport.com/?p=16877</guid>

					<description><![CDATA[<p>Last week, Elizabeth Lopatto published an insightful article in The Verge. It boasted an intriguing title: ​“Silicon Valley has forgotten what normal people want.”​ “Within ... <a title="Who Asked For This?" class="read-more" href="https://calnewport.com/who-asked-for-this/" aria-label="Read more about Who Asked For This?">Read more</a></p>
<p>The post <a href="https://calnewport.com/who-asked-for-this/">Who Asked For This?</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Last week, Elizabeth Lopatto published an insightful article in <em>The Verge</em>. It boasted an intriguing title: <a href="https://www.theverge.com/tldr/915176/nft-metaverse-ai-weirdos">​“Silicon Valley has forgotten what normal people want.”​</a></p>



<p>“Within recent memory, people who made software and hardware understood their job was to serve their customers. It was to identify a need, and then fill it,” she writes. “But at some point following the financial crisis, would-be entrepreneurs got it into their heads that their job was to invent the future, and consumers’ job was to go along with that invented future.”</p>



<p>I certainly noticed this shift when it first began emerging. See, for example, my 2015 article titled, <a href="https://calnewport.com/its-not-your-job-to-figure-out-why-an-apple-watch-might-be-useful/">​“It’s Not Your Job to Figure Out Why an Apple Watch Might Be Useful.”​</a> But it really picked up speed in the last half-decade. Here’s Lopatto with a needle-sharp summary of our current status quo:</p>



<span id="more-16877"></span>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p>“In the place of problem-solving technology, companies have jumped on successive bandwagons like NFTs, the metaverse, and large language models. What these all have in common is that they are not built to really solve a market problem. They are built to make VCs and companies rich.”</p>
</blockquote>



<p>Of these three examples, large language models clearly have the most potential utility. But this doesn’t let AI companies off the hook when it comes to figuring out and communicating those uses.</p>



<p>As Lopatto points out: “Normal people aren’t running around like chickens with their heads cut off, trying to automate every single part of their lives.“ Their biggest exposure to AI is using a tool like ChatGPT as a more verbose Google, or perhaps occasionally formatting an event itinerary. This is cool, and even useful, but at the moment it is probably less positively impactful in their lives than, say, the arrival of the iPod in the early 2000s.</p>



<p>But unlike an iPod, these same ordinary users are forced to hear about AI <em>constantly</em>; not just enthusiast tech bro nonsense, but dark, disturbing, relentless accounts about how everything is about to change in terrible ways that they can’t control.</p>



<p><em>This isn’t sustainable.</em></p>



<p>Generative AI has no shortage of ways that it might, with care, be shaped into genuinely useful products, but this shaping needs to actually happen before the hyper-scalers earn the right to continually harass the psyche of billions of people with breathless pronouncements. Most people don’t care that GPT 5.5, released late last week, underperformed Opus 4.7 on SWE-Bench Pro. They want the AI companies to let them know when they have a product that will actually and notably improve their lives, and until then, they want these companies to leave them alone and try their best not to <a href="https://www.newsweek.com/ai-skeptic-ed-zitron-says-math-on-data-centers-doesnt-add-up-11594219">​crash the economy​</a>.</p>



<p>As Lopatto concludes: “At some point, our Silicon Valley overlords forgot that in order for their vision of the future to be adopted, people had to want it.” They still have a lot of work to do.</p>



<figure class="wp-block-image"><img decoding="async" src="https://embed.filekitcdn.com/e/ekndSb6aixDTy6CAJEGkrv/pqwRyXLTHodBmNohEhe4Yo" alt=""/></figure>



<h3 class="wp-block-heading"><strong>AI Is Destroying the Job Market. Also, AI Is Saving the Job Market</strong></h3>



<p><em>I couldn’t help but add a quick additional note about AI to this week’s newsletter…</em></p>



<p>One of the big stories of the last year was the shrinking post-pandemic job market for recent college graduates. Many media outlets confidently offered an explanation for this shift: AI was automating the work of entry-level positions.</p>



<p>An <a href="https://www.wsj.com/lifestyle/careers/ai-entry-level-jobs-graduates-b224d624">​article​</a> from last summer proclaimed that “AI is wrecking an already fragile job market for college graduates,” going on to note that “ChatGPT and other bots can do many of [the] chores” that used to be handled by entry-level workers. Another <a href="https://www.theguardian.com/us-news/2026/apr/12/college-graduates-job-market-ai">​article​</a>, published only two weeks ago, offered a stark warning: “college graduates can’t find entry-level roles in shrinking market amid rise of AI.”</p>



<p>But then, last week, new job numbers revealed that the entry-level job market for college graduates was rebounding, and hiring in this demographic is now projected to rise significantly. <em>Whoops.</em> I guess AI wasn’t actually automating those jobs. (<a href="https://www.youtube.com/watch?v=XygyGzgrL48">​I told you so​</a>.)</p>



<p>Does this mean the media will stop trying to force this technology into these more routine workforce narratives? If only wishing made it so. A recent <em>Wall Street Journal </em><a href="https://www.wsj.com/lifestyle/careers/are-college-graduates-finally-catching-a-break-in-this-job-market-38c37541">​article​</a> describing these positive numbers included the following line: “In some cases, artificial intelligence is spurring hires by enabling companies to expand services and product lines.”</p>



<p>So, let’s get this straight: AI is simultaneously <em>contracting</em> the job market for recent college graduates while also <em>expanding</em> the job market for recent college graduates.</p>



<p>Is there anything AI can’t do?</p>



<p></p>
<p>The post <a href="https://calnewport.com/who-asked-for-this/">Who Asked For This?</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://calnewport.com/who-asked-for-this/feed/</wfw:commentRss>
			<slash:comments>3</slash:comments>
		
		
			</item>
		<item>
		<title>Brandon Sanderson vs. AI Art</title>
		<link>https://calnewport.com/brandon-sanderson-vs-ai-art/</link>
					<comments>https://calnewport.com/brandon-sanderson-vs-ai-art/#comments</comments>
		
		<dc:creator><![CDATA[Study Hacks]]></dc:creator>
		<pubDate>Mon, 20 Apr 2026 10:00:00 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<guid isPermaLink="false">https://calnewport.com/?p=16872</guid>

					<description><![CDATA[<p>Late last year, the fantasy novelist Brandon Sanderson gave a talk at Dragonsteel Nexus, an annual conference organized by his media company. It was titled, ... <a title="Brandon Sanderson vs. AI Art" class="read-more" href="https://calnewport.com/brandon-sanderson-vs-ai-art/" aria-label="Read more about Brandon Sanderson vs. AI Art">Read more</a></p>
<p>The post <a href="https://calnewport.com/brandon-sanderson-vs-ai-art/">Brandon Sanderson vs. AI Art</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Late last year, the fantasy novelist Brandon Sanderson gave a talk at Dragonsteel Nexus, an annual conference organized by his media company. It was titled, <a href="https://youtu.be/mb3uK-_QkOo?si=evm1LnMf1TCQ5bH6">​“The Hidden Cost of AI Art.”​</a></p>



<p>As Sanderson explains, early in his address: “The surge of large language models and generative AI raises questions that are fascinating, and even if I dislike how the movement is going in relation to writing and art, I want to learn from the experience of what’s happening.”</p>



<p>Sanderson makes it clear that he disapproves of AI-generated art (“my stomach turns”), but he wants to understand better why this is the case. To do so, he begins considering and then ultimately dismissing a series of common objections:</p>



<ul class="wp-block-list">
<li><strong>Does he dislike AI art because of the economic and environmental impacts?</strong> “Well, those do concern me, but if I’m answering honestly, I would still have a problem with it even if AI were not so resource hungry.”</li>



<li><strong>Does he dislike AI art because it’s trained on the work of existing artists?</strong> “ Well, I don’t like that. But even if it were trained using no copyrighted work, I’d still be concerned.”</li>



<li><strong>Does he just hate the idea of a machine replacing a person?</strong> Sanderson references the folk tale of John Henry attempting to beat a steam drill in a tunnel-digging competition that culminates in Henry’s death. “We respect him, but as a society we chose the steam drill. And I would too&#8230;The truth is, I’m more than happy to have steam engines drilling tunnels for me to drive through.”</li>
</ul>



<p><em>So what is it?</em></p>



<span id="more-16872"></span>



<p>Sanderson ultimately lands on a more personal reason. Talking about his struggles with his first (failed) book manuscripts, he identifies the key value of art: it changes the artist who attempts it. As he elaborates:</p>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p>“Maybe someday the language models will be able to write books better than I can. But here’s the thing: Using those models in such a way absolutely misses the point, because it looks at art only as a product. Why did I write [my first manuscript]?&#8230; It was for the satisfaction of having written a novel, feeling the accomplishment, and learning how to do it. I tell you right now, if you’ve never finished a project on this level, it’s one of the most sweet, beautiful, and transcendent moments. I was holding that manuscript, thinking to myself, ‘I did it. I did it.’”</p>
</blockquote>



<p>As a writer myself, I’ve also been thinking about this question recently. I like Sanderson’s take, but I’ve been developing one of my own. I understand art to be an act of deep human communication, in which the artist uses a tangible medium, such as a page of prose or a painted canvas, to transmit a complex internal cognitive state from their brain to that of their audience.</p>



<p>It’s telepathy. And it’s one of the most beautiful and human things we do.</p>



<p>This makes the idea of reading a book written by a language model, or watching a film generated by a prompt, intrinsically absurd, if not anti-human. It’s the heroin needle providing a quixotic simulation of love.</p>



<p>What really struck me about Sanderson’s talk, however, was his conclusion. If art is deeply human, he argues, then it’s up to us to define it. “That’s the great thing about art – we define it, and we give it meaning,” he says. “The machines can spit out manuscript after manuscript after manuscript. They can pile them to the pillars of heaven itself. But all we have to do is say ‘no.’”</p>



<p>I’ve noticed a trend in recent AI commentary toward a certain nihilistic passivity. You probably know what I&#8217;m talking about – the now popular style of essay in which the author, with a sort of worldly weariness, lays out some grim scenario in which AI destroys something sacred, and then sort of just leaves it there, like a cat dropping a dead bird on the doorstep.</p>



<p>I’m getting tired of this meekness.</p>



<p>Sanderson reminds us that we have agency. In the areas that matter most, it’s us, not the whims of Sam Altman or Dario Amodei, that determine how we shape our existence. All we have to do is say “no.”</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h3 class="wp-block-heading">Correction: </h3>



<p>In last week&#8217;s&nbsp;<a href="https://preview.convertkit-mail4.com/click/dpheh0hzhm/aHR0cHM6Ly93d3cueW91dHViZS5jb20vd2F0Y2g_dj1rLThzdFFDZVFpRQ==" target="_blank" rel="noreferrer noopener">AI Reality Check episode</a>&nbsp;of my podcast, I said the following:</p>



<p>&#8220;If you go back and look at the release notes for Anthropic&#8217;s earlier, less powerful opus 4.6 LLM, they say the following: their researchers used Opus to find, quote, &#8216;over 500 exploitable zero-day vulnerabilities, some of which are decades old.&#8217; And let&#8217;s stop for a moment because that note, which was hidden in the system card for opus 4.6, is almost word for word what anthropic said about Mythos.&#8221;</p>



<p>Some of this wording was sloppy, so I want to clarify it here. I was referring to&nbsp;<a href="https://preview.convertkit-mail4.com/click/dpheh0hzhm/aHR0cHM6Ly9yZWQuYW50aHJvcGljLmNvbS8yMDI2L3plcm8tZGF5cy8=" target="_blank" rel="noreferrer noopener">this report</a>&nbsp;on Opus 4.6, which Anthropic published the same day it was released. This is not technically the system card for Opus 4.6, but it is accurately described as&nbsp;<em>release notes</em>&nbsp;(or perhaps&nbsp;<em>supplementary release notes</em>).</p>



<p>This report said: &#8220;Opus 4.6 found high-severity vulnerabilities, some that had gone undetected for decades.&#8221; In another place, it said: &#8220;So far, we&#8217;ve found and validated more than 500 high-severity vulnerabilities.&#8221; Both the title of the report and the conclusion refer to these vulnerabilities as “0-day.”</p>



<p>The specific quote I provided, however, does not appear in the report. It&#8217;s actually a summary of the report from<a href="https://preview.convertkit-mail4.com/click/dpheh0hzhm/aHR0cHM6Ly94LmNvbS9fRGFuaWVsU2luY2xhaXIvc3RhdHVzLzIwMTk1MjcxMDk4ODczNzc1NTc=" target="_blank" rel="noreferrer noopener">&nbsp;this tweet</a>. In my opinion, the summary is accurate, but the way I worded the above implies that it was actually found in the report, which it was not.</p>



<p><em>Thank you to the AI researcher who pointed out these issues. I appreciate corrections! You can always send concerns or notes to podcast@calnewport.com.</em></p>
<p>The post <a href="https://calnewport.com/brandon-sanderson-vs-ai-art/">Brandon Sanderson vs. AI Art</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://calnewport.com/brandon-sanderson-vs-ai-art/feed/</wfw:commentRss>
			<slash:comments>8</slash:comments>
		
		
			</item>
		<item>
		<title>Is Claude Mythos “Terrifying” or Just Hype?</title>
		<link>https://calnewport.com/is-claude-mythos-terrifying-or-just-hype/</link>
					<comments>https://calnewport.com/is-claude-mythos-terrifying-or-just-hype/#comments</comments>
		
		<dc:creator><![CDATA[Study Hacks]]></dc:creator>
		<pubDate>Mon, 13 Apr 2026 10:00:00 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<guid isPermaLink="false">https://calnewport.com/?p=16865</guid>

					<description><![CDATA[<p>Last week, millions of New York Times readers were subjected to ​an alarming column​ by Thomas Friedman. “Normally right now I would be writing about ... <a title="Is Claude Mythos “Terrifying” or Just Hype?" class="read-more" href="https://calnewport.com/is-claude-mythos-terrifying-or-just-hype/" aria-label="Read more about Is Claude Mythos “Terrifying” or Just Hype?">Read more</a></p>
<p>The post <a href="https://calnewport.com/is-claude-mythos-terrifying-or-just-hype/">Is Claude Mythos “Terrifying” or Just Hype?</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Last week, millions of <em>New York Times</em> readers were subjected to <a href="https://www.nytimes.com/2026/04/07/opinion/anthropic-ai-claude-mythos.html">​an alarming column​</a> by Thomas Friedman. “Normally right now I would be writing about the geopolitical implications of the war with Iran,” Friedman begins, before soon continuing, “but I want to interrupt that thought to highlight a stunning advance in artificial intelligence — one that arrived sooner than expected and that will have equally profound geopolitical implications.”</p>



<p>The “stunning advance” was the release of Anthropic&#8217;s new LLM, named Claude Mythos. In a lengthy <a href="https://www.anthropic.com/glasswing">​press release​</a>, Anthropic announced that the model would be made available to a consortium of business partners, but not to the general public. To justify this decision, Anthropic cited their concerns about its effectiveness at finding security vulnerabilities in source code, noting: “AI models have reached a level of coding capability where they can surpass all but the most skilled humans at finding and exploiting software vulnerabilities.”</p>



<p>They go on to explain that Mythos “has already found thousands of high-severity vulnerabilities, including some in <em>every major operating system and web browser</em>.<em>”</em></p>



<p>This announcement clearly rattled Friedman, who called Anthropic’s decision not to release the model a “terrifying warning sign,” writing:</p>



<p>“Holy cow! Superintelligent A.I. is arriving faster than anticipated, at least in this area…If this A.I. tool were, indeed, to become widely available, it would mean the ability to hack any major infrastructure system — a hard and expensive effort that was once essentially the province only of private-sector experts and intelligence organizations — will be available to every criminal actor, terrorist organization and country, no matter how small.”</p>



<p>Friedman was far from alone in this concern. Many major news outlets expressed similar unease about this scary new development, including <a href="https://finance.yahoo.com/video/is-anthropics-claude-mythos-an-ai-nightmare-waiting-to-happen-203000700.html">​one particularly anxiety-provoking headline​</a> that asked if Mythos was an “AI nightmare waiting to happen?”</p>



<p>So, what’s really going on here?</p>



<p>I thought it was worth taking a moment to look closer, not just to address the specific worries about Mythos, but also to help recalibrate, more generally, how those of us seeking depth in a distracted world should consume AI news.</p>



<p class="has-text-align-center">~~~</p>



<p>When I talked to people who were spooked by Friedman’s column, they tended to be under the impression that this ability to find and exploit security vulnerabilities was a new phenomenon; a skill that emerged unexpectedly in Mythos, &#8220;terrifying&#8221; those who studied it.</p>



<p>In reality, security researchers have been worried about using LLMs for this purpose since the beginning of consumer LLMs.</p>



<p>Back in 2024, for example, IBM researchers published <a href="https://arxiv.org/abs/2404.08144">​a splashy study​</a> about using GPT-4 to attack security vulnerabilities. They found that GPT-4 successfully exploited 87% of the vulnerabilities that it was presented, as compared to close to 0% for GPT 3.5. “Our findings raise questions around the widespread deployment of highly capable LLM agents,” they concluded.</p>



<p>To be fair, in the case of GPT-4, researchers were assessing whether an LLM could write code to exploit a known vulnerability. Mythos, however, can also find these vulnerabilities from scratch. But this isn’t new either.</p>



<p>Accompanying the release notes for Anthropic’s earlier Opus 4.6 LLM was <a href="https://www.reddit.com/r/Anthropic/comments/1r05i5g/opus_46_found_over_500_exploitable_0days_some_of/">​the observation​</a> that Anthropic’s security team used the model to find “over 500 exploitable 0-day [vulnerabilities], some of which are decades old.” This is almost word-for-word what Anthropic said last week about Mythos, the main difference being that they replaced 500 with “thousands.”</p>



<p>We are not, therefore, talking about a new capability, but rather one that has been around for multiple years.</p>



<p>The relevant question then becomes, how much better is Mythos at finding vulnerabilities? It’s hard to tell for sure because Anthropic has kept their new model private. They did, however, release that Mythos scored 83.1% on a well-known cybersecurity benchmark. For comparison, Opus 4.6 scored 66.6% on this same test.</p>



<p>In general, benchmark results should be taken with a grain of salt as they represent specific (often narrow) tests that researchers can tune their models to pass. But even if we accept that this particular measure is useful, a sixteen percentage point increase seems to represent solid incremental progress more than a nightmarish leap.</p>



<p>When we turn our attention to actual results, the waters become even murkier. In a recent Substack post (<a href="https://garymarcus.substack.com/p/three-reasons-to-think-that-the-claude">​which is worth reading​</a>), Gary Marcus rounds up responses from security researchers who took a closer look at the specific exploits that Anthropic reported that Mythos discovered. They were not impressed.</p>



<ul class="wp-block-list">
<li>Philo Groves, for example, <a href="https://x.com/philogroves/status/2042195139477557499?s=61">​noted​</a> that Mythos’s attention-grabbing attack on the Firefox browser required certain common security features to be disabled, and it built on results previously discovered by Opus. (“Shocker,” he concludes sardonically.)</li>



<li>The CEO of the AI company HuggingFace then <a href="https://x.com/clementdelangue/status/2041953761069793557?s=61">​reported​</a> that they took all of the specific vulnerabilities that Anthropic highlighted and “ran them through small, cheap, open-weight models.” What did they find? “Those models recovered much of the same analysis.”</li>
</ul>



<p>Since Marcus published his essay, I’ve come across several more similar findings:</p>



<ul class="wp-block-list">
<li>The AI security expert Stanislav Fort ran <a href="https://x.com/stanislavfort/status/2041922370206654879">​an experiment​</a> to see if existing, cheap open-weight models could find the same vulnerability in FreeBSD (an open-source operating system) that Anthropic touted as evidence of Mythos’s scary abilities to uncover bugs that had been hiding for decades. The result: all eight existing models they tested discovered the same issue.</li>



<li>Meanwhile, the renowned security researcher Bruce Schneier <a href="https://www.youtube.com/watch?v=PsKVSHjres4">​weighed in​</a>, similarly concluding: “You don’t need Mythos to find the vulnerabilities they found.”</li>
</ul>



<p>And of course, it doesn’t help that a week before Anthropic released this supposedly super-powered vulnerability detector, they accidentally leaked the Claude Code source, and security researchers immediately found <a href="https://www.securityweek.com/critical-vulnerability-in-claude-code-emerges-days-after-source-leak/">​serious vulnerabilities​</a>. (I guess Anthropic forgot to use Mythos to clean up their own software…)</p>



<p class="has-text-align-center">~~~</p>



<p>What’s really happening?</p>



<p>It’s fair to say that LLMs have created <em>significant</em> cybersecurity concerns that researchers have been scrambling to address in recent years. It’s also fair to say, however, that we don’t yet have evidence that Claude Mythos significantly changed this reality. If anything, some of the early independent testing by security researchers implies that Mythos might be better understood as a version of Opus 4.6 tuned to perform better on a handful of benchmarks. And yet, many still took Anthropic at their word and covered this model’s release as a catastrophic event.</p>



<p>In a <a href="https://www.youtube.com/watch?v=mcN1VTTIjQs">​recent video​</a>, the AI commentator Mo Bitar compared Anthropic’s model rollouts to Apple iPhone launches, where every year they resell you the same product with minor improvements. “Except here,” he adds, “the product is existential dread.”</p>



<p>And we keep falling for it.</p>



<p>I think we’ve entered a stage where we need to almost entirely discount any claims made by the AI companies themselves <em>until</em> we can independently verify what’s actually going on.</p>
<p>The post <a href="https://calnewport.com/is-claude-mythos-terrifying-or-just-hype/">Is Claude Mythos “Terrifying” or Just Hype?</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://calnewport.com/is-claude-mythos-terrifying-or-just-hype/feed/</wfw:commentRss>
			<slash:comments>3</slash:comments>
		
		
			</item>
		<item>
		<title>In Defense of Thinking</title>
		<link>https://calnewport.com/in-defense-of-thinking-2/</link>
					<comments>https://calnewport.com/in-defense-of-thinking-2/#comments</comments>
		
		<dc:creator><![CDATA[Study Hacks]]></dc:creator>
		<pubDate>Mon, 30 Mar 2026 10:00:00 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<guid isPermaLink="false">https://calnewport.com/?p=16857</guid>

					<description><![CDATA[<p>Ten years ago, I published ​Deep Work​. It was my second mainstream hardcover idea book. The previous title, ​So Good They Can’t Ignore You​, hadn’t ... <a title="In Defense of Thinking" class="read-more" href="https://calnewport.com/in-defense-of-thinking-2/" aria-label="Read more about In Defense of Thinking">Read more</a></p>
<p>The post <a href="https://calnewport.com/in-defense-of-thinking-2/">In Defense of Thinking</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Ten years ago, I published <a href="https://www.amazon.com/Deep-Work-Focused-Success-Distracted/dp/1455586692">​<em>Deep Work</em>​</a><em>. </em>It was my second mainstream hardcover idea book. The previous title, <a href="https://www.amazon.com/dp/1455509124/">​<em>So Good They Can’t Ignore You</em>​</a><em>,</em> hadn’t sold as well as we hoped, so the expectations were lower for this follow-up.</p>



<p>This turned out to be freeing, as it allowed me to write <em>Deep Work</em> largely for myself – exploring the conceptual edges of the issues surrounding distraction that interested me most.</p>



<p>I was fascinated, for example, by the economic reality that so many knowledge work organizations systematically undervalued focus, and was convinced that this provided a massive opportunity for those willing to correct for this mistake. In this way, I saw myself as articulating something like <em>Moneyball</em> for the cubicle class. I also firmly believed that the act of thinking was at the core of the post-Paleolithic human experience; the source of our greatest ideas, satisfactions, and even moments of transcendence.</p>



<p>This mixture of the economic and philosophical was different from the typical book in this genre at the time. Readers probably expected that I would open on a breathless tale of an overworked executive, then regurgitate some stats about interruptions, before proceeding with long lists of tips calibrated to be practical, but also not too challenging, presented in a conversational tone and accompanied by clearly manipulated case studies.</p>



<p>But <em>Deep Work</em> was much weirder and more intense than that. Re-reading it recently, I was struck by how many of my stories had nothing to do with the knowledge sector at all. I quoted philosophers of religion and a blacksmith who forged swords with ancient techniques. I profiled a memory champion and discussed <em>chavruta,</em> the Jewish practice of studying Talmud or Torah in pairs. Rather than opening the book on a frustrated executive, I focused on Carl Jung’s efforts to break free from Sigmund Freud’s capriciousness. It was a direct look at the sources and ideas that most resonated with me.</p>



<p>This idiosyncratic approach seemed to reveal something fundamentally true about the problematic state of work at that time, as the book soon found an audience, going on to sell more than two million copies in over forty-five languages. (In its wake, <em>So Good They Can’t Ignore You</em> finally found its groove as well, quietly selling more than half a million copies, providing me with a dash of retrospective vindication.)</p>



<p>All of this led me recently to ask a natural follow-up question: <strong>How have things changed since that book first came out in 2016?</strong></p>



<p>I tackled this query in <a href="https://www.nytimes.com/2026/03/27/opinion/technology-mental-fitness-cognitive.html">​a long-form essay​</a> I published in the <em>New York Times</em> over the weekend. My answer wasn’t optimistic:</p>



<span id="more-16857"></span>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p>“The problems I focused on in <em>Deep Work</em>, and in my writing since, have been getting steadily worse. In 2016 my main concern was helping people find enough free time for deep work. Today I think we’re rapidly losing the ability to think deeply at all, regardless of how much space we can find in our schedules for these efforts.&#8221;</p>
</blockquote>



<p>Distractions in the workplace intensified over the past decade with the addition of instant messaging tools like Slack and low-friction digital meeting programs like Zoom. Outside of work, social media, which was generally still admired when <em>Deep Work</em> came out, has morphed into an addictive TikTok-ified slurry of optimized brain rot. Meanwhile, new AI tools offer quick-fix short-cuts to whatever intellectually engaging work activities remain.</p>



<p>None of this is great news.</p>



<p>So, what should we do? The obvious short answer is to read <a href="https://www.amazon.com/Deep-Work-Focused-Success-Distracted/dp/1455586692">​<em>Deep Work</em>.​</a> (Or, if you already have, buy some copies for people you know who need to hear its message!)</p>



<p>But that’s only a small step toward our larger goal of a world in which we once again respect the act of cognition. In my <em>Times</em> piece, I suggest a louder response: we launch a revolution in defense of thinking.</p>



<p>I go on to suggest multiple concrete actions that such a revolution can include, such as:</p>



<ul class="wp-block-list">
<li>Stop consuming social media (which is, if we are being honest, digital junk food and something adults largely need to eliminate from a healthy content diet).</li>



<li>Keep your phone plugged in and charging when at home instead of on your person.</li>



<li>Push Congress to follow Australia&#8217;s lead and ban social media for kids.</li>



<li>Build work cultures in which phones and laptops stay out of meetings, and find collaboration strategies that don’t require constant messaging.</li>



<li>Stop vague demands to “use AI” and instead carefully integrate these tools where they actually make us smarter, not just busier.</li>
</ul>



<p>But more important than any specific suggestion is the larger spirit of revolution. “I’m done ceding my brain — the core of all that makes me who I am — to the financial interests of a small number of technology billionaires or the shortsighted conveniences of hyperactive communication styles,” I write in the conclusion of my <em>Times</em> op-ed. “It’s time to move past fretting about our slide into the cognitive shallows and decide to actually do something about it.”</p>
<p>The post <a href="https://calnewport.com/in-defense-of-thinking-2/">In Defense of Thinking</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://calnewport.com/in-defense-of-thinking-2/feed/</wfw:commentRss>
			<slash:comments>23</slash:comments>
		
		
			</item>
		<item>
		<title>Avoiding Digital Productivity Traps</title>
		<link>https://calnewport.com/avoiding-digital-productivity-traps/</link>
					<comments>https://calnewport.com/avoiding-digital-productivity-traps/#comments</comments>
		
		<dc:creator><![CDATA[Study Hacks]]></dc:creator>
		<pubDate>Mon, 23 Mar 2026 10:00:00 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<guid isPermaLink="false">https://calnewport.com/?p=16818</guid>

					<description><![CDATA[<p>​Last week​ in this newsletter, I summarized some interesting results from ​a study​ that analyzed the behavior of 164,000 knowledge workers. It found that introducing ... <a title="Avoiding Digital Productivity Traps" class="read-more" href="https://calnewport.com/avoiding-digital-productivity-traps/" aria-label="Read more about Avoiding Digital Productivity Traps">Read more</a></p>
<p>The post <a href="https://calnewport.com/avoiding-digital-productivity-traps/">Avoiding Digital Productivity Traps</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p><a href="https://calnewport.com/why-hasnt-ai-made-work-easier/">​Last week​</a> in this newsletter, I summarized some interesting results from <a href="https://www.wsj.com/tech/ai/ai-isnt-lightening-workloads-its-making-them-more-intense-e417dd2c">​a study​</a> that analyzed the behavior of 164,000 knowledge workers. It found that introducing AI tools increased administrative tasks by more than 90% while reducing deep work effort by almost 10%.</p>



<p>The problem, I concluded, was that digital productivity tools sometimes speed up the <em>wrong</em> tasks, which might feel efficient in the moment, but lead us to accomplish less over time. As I emphasized, AI is not the only technology to produce this paradoxical side effect —we saw something similar with email, mobile computing, and online meeting software as well.</p>



<p><em>So, what’s the solution to avoid these traps?</em></p>



<p>In <a href="https://open.spotify.com/show/0e9lFr3AdJByoBpM6tAbxD?si=c92344b6836b4c76">​today’s episode​</a> of my podcast, I suggested three ideas that might help. I want to summarize them here as well:</p>



<span id="more-16818"></span>



<p><strong>Idea #1:</strong> Use a Better Scoreboard</p>



<p>Make sure you measure what <em>actually</em> matters in your job. If you’re a professor at a research institution, for example, this might be the number of papers you publish per year. If you’re a team manager, it might be the number of priority projects completed per month.</p>



<p>When you introduce new digital productivity tools into your workflow, don’t focus too much on their impact on individual tasks (e.g., “Wow! That email was much faster to send than a fax,” or “AI just finished a task in 20 minutes that would have taken me 3 hours!”). Pay attention instead to your scoreboard. If you’re not producing more valuable output than before, the tool isn’t really making you more productive.</p>



<p><strong>Idea #2: </strong>Focus on the Right Bottlenecks</p>



<p>If you look closer at many knowledge work projects, you’ll identify a key <em>bottleneck</em> that determines how fast they can be accomplished. If you want to become more productive, you should look for ways to deploy tools that improve this specific step.</p>



<p>When working on <em>Deep Work</em>, for example, I spoke with a prominent Wharton professor who told me that one of the keys to publishing journal papers in his field was access to interesting data sets. He published more papers per year than most of his peers, largely because he spent more time building relationships with companies and institutions in search of good data. This was the bottleneck for his work.</p>



<p>Accordingly, any tool that could help him cultivate more such relationships and gather better data from the relationships he had already formed would directly improve his productivity. Compare this, for example, to using Claude Code to speed up the process of producing plots for his papers. This might, in limited windows of time, make his job more convenient, but not necessarily increase the number of papers he publishes per year.</p>



<p><strong>Idea #3:</strong> Separate Deep from Shallow Work</p>



<p>My final idea is the simplest: on your daily calendar, clearly separate time for focused effort that directly produces value from administrative, logistical, and collaborative tasks. In this way, if a digital productivity tool ends up accidentally increasing the volume of shallow work you face each day, you’ll limit the damage to your ability to make progress on important projects.</p>



<p>This makes it easier to experiment with different tools without worrying that you might end up — like many of the subjects in the study cited above — suddenly overwhelmed by the ultra-fast processing of minutiae while the big things slowly languish.</p>
<p>The post <a href="https://calnewport.com/avoiding-digital-productivity-traps/">Avoiding Digital Productivity Traps</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://calnewport.com/avoiding-digital-productivity-traps/feed/</wfw:commentRss>
			<slash:comments>3</slash:comments>
		
		
			</item>
		<item>
		<title>Why Hasn’t AI Made Work Easier?</title>
		<link>https://calnewport.com/why-hasnt-ai-made-work-easier/</link>
					<comments>https://calnewport.com/why-hasnt-ai-made-work-easier/#comments</comments>
		
		<dc:creator><![CDATA[Study Hacks]]></dc:creator>
		<pubDate>Mon, 16 Mar 2026 10:00:00 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<guid isPermaLink="false">https://calnewport.com/?p=16813</guid>

					<description><![CDATA[<p>I’ve been studying the intersection of digital technology and office work for quite some time. (I find it hard to believe that my book, ​Deep ... <a title="Why Hasn’t AI Made Work Easier?" class="read-more" href="https://calnewport.com/why-hasnt-ai-made-work-easier/" aria-label="Read more about Why Hasn’t AI Made Work Easier?">Read more</a></p>
<p>The post <a href="https://calnewport.com/why-hasnt-ai-made-work-easier/">Why Hasn’t AI Made Work Easier?</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>I’ve been studying the intersection of digital technology and office work for quite some time. (I find it hard to believe that my book, <a href="https://www.amazon.com/Deep-Work-Focused-Success-Distracted/dp/1455586692">​<em>Deep Work</em>​</a>, just passed its ten-year anniversary!?) Here’s a pattern I’ve observed again and again:</p>



<ul class="wp-block-list">
<li>A new technology promises to speed up some annoying aspects of our jobs.</li>



<li>Everyone gets excited about freeing up more time for deep work and leisure.</li>



<li>We end up <em>busier</em> than before without producing more of the high-value output that actually moves the needle.</li>
</ul>



<p>This happened with the front-office IT revolution, and email, and mobile computing, and once again with video-conferencing.</p>



<p>I’m now starting to fear that we’re beginning to encounter the same thing with AI as well.</p>



<p>My worries were stoked, in part, by a recent article in the <em>Wall Street Journal</em>, titled <a href="https://www.wsj.com/tech/ai/ai-isnt-lightening-workloads-its-making-them-more-intense-e417dd2c">​“AI Isn’t Lightening Workloads. It’s Making Them More Intense.”​</a></p>



<p>The piece cites new research from the software company ActivTrak, which analyzed the digital activity of 164,000 workers across more than 1,000 employers. What makes the study notable is its methodology: it tracked individual AI users for 180 days before and after they began using these tools, providing clear insight into what changed. The results?</p>



<p>“ActivTrak found AI intensified activity across nearly every category: The time they spent on email, messaging and chat apps more than doubled, while their use of business-management tools, such as human-resources or accounting software, rose 94%.“</p>



<p>The one category where activity was <em>not</em> intensified, however, was deep work:</p>



<p>“[T]he amount of time AI users devoted to focused, uninterrupted work—the kind of concentration often required for figuring out complex problems, writing formulas, creating and strategizing—fell 9%, compared with nearly no change for nonusers.”</p>



<p>This is a worst-case scenario: you work faster and harder, but mainly on shallow, mentally taxing tasks (because of all the context shifting they require) that only indirectly help the bottom line compared to harder efforts.</p>



<span id="more-16813"></span>



<p>It’s not quite clear why AI tools are having this impact. One tantalizing clue, however, comes from Berkeley professor Aruna Ranganathan, who is quoted in the article saying: “AI makes additional tasks feel easy and accessible, creating a sense of momentum.”</p>



<p>This points toward a pattern similar to what happened when email first arrived. It was undeniably true that sending emails was more efficient than wrangling fax machines and voicemail. But once workers gained access to low-friction communication, they transformed their days into a furious flurry of back-and-forth messaging that felt “productive” in the <a href="https://www.amazon.com/Slow-Productivity-Accomplishment-Without-Burnout/dp/0593544854/">​abstract, activity-centric sense​</a> of that term, but ultimately hurt almost every other aspect of their jobs and <a href="https://www.newyorker.com/tech/annals-of-technology/e-mail-is-making-us-miserable">​made everyone miserable​</a>.</p>



<p>AI tools might be replicating this dynamic with small, self-contained tasks. Users are now furiously bouncing ideas back and forth with chatbots, iteratively refining text and generating drafts of memos and slide decks that are often <a href="https://hbr.org/2025/09/ai-generated-workslop-is-destroying-productivity">​too sloppy ​</a>to be useful. If they’re particularly tech savvy, perhaps they’re even monitoring the efforts of agent swarms deployed to parallelize such efforts even further. Once again, this all seems “productive” in the sense that these individual tasks appear to be happening faster, and activity seems intensified overall.</p>



<p>But are we sure we’re accelerating the right parts of our jobs?</p>



<hr class="wp-block-separator has-text-color has-global-color-8-color has-alpha-channel-opacity has-global-color-8-background-color has-background is-style-default"/>



<h3 class="wp-block-heading"><strong>I Need Your Help</strong></h3>



<p>I’m working on an article for a major publication about the move toward simple, high-friction, single-use technologies like the <a href="https://tincan.kids/">​Tin Can phone​</a>. If you have a Tin Can phone/are on the waiting list, or have recently embraced similar retro technologies, and are willing to talk, please send me an email at <a href="mailto:podcast@calnewport.com">​<strong>podcast@calnewport.com</strong>​</a>. I want to hear about your motivations and experience!</p>



<figure class="wp-block-image"><img fetchpriority="high" decoding="async" width="2400" height="240" src="https://calnewport.com/wp-content/uploads/2026/03/pqwRyXLTHodBmNohEhe4Yo.png" alt="" class="wp-image-16815" srcset="https://calnewport.com/wp-content/uploads/2026/03/pqwRyXLTHodBmNohEhe4Yo.png 2400w, https://calnewport.com/wp-content/uploads/2026/03/pqwRyXLTHodBmNohEhe4Yo-300x30.png 300w, https://calnewport.com/wp-content/uploads/2026/03/pqwRyXLTHodBmNohEhe4Yo-1024x102.png 1024w, https://calnewport.com/wp-content/uploads/2026/03/pqwRyXLTHodBmNohEhe4Yo-768x77.png 768w, https://calnewport.com/wp-content/uploads/2026/03/pqwRyXLTHodBmNohEhe4Yo-1536x154.png 1536w, https://calnewport.com/wp-content/uploads/2026/03/pqwRyXLTHodBmNohEhe4Yo-2048x205.png 2048w" sizes="(max-width: 2400px) 100vw, 2400px" /></figure>



<h3 class="wp-block-heading"><strong>AI Reality Check</strong>: Is Claude Conscious?</h3>



<p>If you were following AI news last week, you might have noticed a barrage of concerning headlines about Anthropic’s Claude LLM, including:</p>



<ul class="wp-block-list">
<li><a href="https://futurism.com/artificial-intelligence/anthropic-ceo-unsure-claude-conscious">​“Anthropic CEO Says Company No Longer Sure Whether Claude is Conscious.”​</a></li>



<li><a href="https://www.newsnationnow.com/jesse-weber-live/claude-ai-consciousness/">​“Is AI Assistant Claude Conscious – and Suffering from Anxiety?”​</a></li>



<li><a href="https://www.ndtv.com/world-news/is-claude-conscious-anthropic-ceo-dario-amodei-says-possibility-cant-be-ruled-out-11175771">​“Is Claude Conscious? Anthropic CEO Says Possibility Can’t Be Ruled Out”​</a></li>
</ul>



<p><em>Here’s what happened.</em> Anthropic infamously puts outlandish warnings and observations in their release notes for their new models because, I suppose, they think it makes them look more safety-aware and responsible (e.g., their classic <a href="https://www.aipanic.news/p/ai-blackmail-fact-checking-a-misleading">​AI blackmail farce​</a>).</p>



<p>True to form, in the notes accompanying the recent release of Opus 4.6, they wrote that the model <strong>“expresses occasional discomfort with the experience of being a product</strong>” and would <strong>“assign itself a 15 to 20 percent probability of being conscious under a variety of prompting circumstances.”</strong></p>



<p>That last part is key. With the right prompts, you can induce an LLM to describe itself as anything you want. Remember: the goal of LLMs is to complete whatever story they’re provided as input. If you wind a model up – even subtly – to write a story from the perspective of being a conscious AI, it will oblige.</p>



<p>Anyway, in <a href="https://www.nytimes.com/2026/02/12/opinion/artificial-intelligence-anthropic-amodei.html">​a recent interview​</a>, Ross Douthat asked Anthropic CEO Dario Amodei about this particular release note. Amodei answered, in part, by saying:</p>



<p>“We don’t know if the models are conscious. We are not even sure that we know what it would mean for a model to be conscious or whether a model can be conscious. But we’re open to the idea that it could be.”</p>



<p>Of course, you could say the same thing about a vacuum cleaner. It’s a non-answer containing no actual information or testable claims. But, the internet being the internet, ran with it. <em>Sigh.</em></p>
<p>The post <a href="https://calnewport.com/why-hasnt-ai-made-work-easier/">Why Hasn’t AI Made Work Easier?</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://calnewport.com/why-hasnt-ai-made-work-easier/feed/</wfw:commentRss>
			<slash:comments>5</slash:comments>
		
		
			</item>
		<item>
		<title>The Original Attention Crisis</title>
		<link>https://calnewport.com/the-original-attention-crisis/</link>
					<comments>https://calnewport.com/the-original-attention-crisis/#comments</comments>
		
		<dc:creator><![CDATA[Study Hacks]]></dc:creator>
		<pubDate>Mon, 09 Mar 2026 10:00:00 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<guid isPermaLink="false">https://calnewport.com/?p=16810</guid>

					<description><![CDATA[<p>I recently heard from a historian of science at All Souls College, Oxford. He forwarded me ​an essay​ he wrote about Nicolaus Steno, a seventeenth-century ... <a title="The Original Attention Crisis" class="read-more" href="https://calnewport.com/the-original-attention-crisis/" aria-label="Read more about The Original Attention Crisis">Read more</a></p>
<p>The post <a href="https://calnewport.com/the-original-attention-crisis/">The Original Attention Crisis</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>I recently heard from a historian of science at All Souls College, Oxford. He forwarded me <a href="https://nunocastelbranco.substack.com/p/focused-work-in-early-modern-times">​an essay​</a> he wrote about Nicolaus Steno, a seventeenth-century anatomist and geologist who was later ordained as a Catholic Bishop.</p>



<p>Steno’s training as a scholar unfolded in a period challenged by a novel problem: information overload. Here’s how the essay describes it:</p>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p>“Books were a leading distraction in the early modern period—and how envious we should be of those times. From the 1500s onward, with the development of the printing press and the humanist revival of ancient philosophies, knowledge became available at a much greater pace than ever before.”</p>
</blockquote>



<p>This created pressing questions for aspiring thinkers, including: “How do we decide what to read? How long should we read it for? Must every single chapter be excerpted?”</p>



<p>Part of the solution was the development of “new note-taking techniques,” including the copying of excerpts into a master notebook called a book of commonplaces. (For more on this technique, I recommend William Powell’s delightful 2010 techno-history, <a href="https://www.amazon.com/Hamlets-BlackBerry-Building-Good-Digital/dp/0061687170/">​<em>Hamlet’s Blackberry</em>​</a>).</p>



<p>But as the essay on Steno elaborates, better notes weren’t enough on their own, as there were simply too many good books available. In response to this reality, Steno, during his university studies in the 1650s, innovated some more advanced attention management strategies:</p>



<span id="more-16810"></span>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p>“[H]e learned to focus on specific themes, rather than letting his mind read multiple things quickly. A ‘harmful hastening should be avoided’ as he put it. His solution was to ‘stick to one topic.’</p>



<p>In practice, that meant blocking specific moments of time to go through the hardest tasks. As he wrote in his personal notebook, ‘before noon nothing must be done except medical things.’ … As Steno told a friend, he took ‘almost all the morning hours’ to read the works of the Church Fathers and old biblical manuscripts available at the Medici library.”</p>
</blockquote>



<p>In other words, Steno created a method that combines what we might now call <a href="https://www.amazon.com/Slow-Productivity-Accomplishment-Without-Burnout/dp/0593544854">​slow productivity​</a>, <a href="https://www.amazon.com/Deep-Work-Focused-Success-Distracted/dp/1455586692">​deep work​</a>, and <a href="https://www.timeblockplanner.com/">​time blocking​</a>.</p>



<p>The lessons here are clear. The use of our brains to think deeply about meaningful ideas isn’t new. It’s been at the core of the human experience since the early modern period, when access to sophisticated information first became somewhat widespread.</p>



<p>The best practices developed back then remain the best practices today: avoid overload, focus on one thing at a time, and block off specific hours in your day for your most mentally demanding efforts.</p>



<figure class="wp-block-image"><img decoding="async" src="https://embed.filekitcdn.com/e/ekndSb6aixDTy6CAJEGkrv/pqwRyXLTHodBmNohEhe4Yo" alt=""/></figure>



<h3 class="wp-block-heading"><strong>AI Reality Check</strong>:</h3>



<p>Two weeks ago, a small financial services firm, Citrini Research, published <a href="https://www.citriniresearch.com/p/2028gic">​an essay​</a> describing a bleak scenario in which AI agents destroy the white-collar job market in the near future. The piece went viral and was <a href="https://www.bloomberg.com/news/articles/2026-02-24/citrini-founder-shocked-his-ai-prediction-spurred-stocks-selloff?embedded-checkout=true">​cited as a factor​</a> in a modest decline of the S&amp;P 500 the next day.</p>



<p>The Citrini essay wasn’t the first to float this scenario. In recent weeks, there have been multiple credulous articles and op-eds in major publications proposing similar outcomes (e.g., <a href="https://www.theatlantic.com/ideas/2026/02/ai-white-collar-jobs/686031/">​1​</a>, <a href="https://www.theatlantic.com/magazine/2026/03/ai-economy-labor-market-transformation/685731/">​2​</a>, <a href="https://www.nytimes.com/2026/03/05/opinion/ai-jobs-white-collar-apocalpyse.html">​and 3​</a>). But the negative impact on the stock market seems to have been the last straw for serious economists who began to push back on these technological ghost stories last week. (I particularly enjoyed a Deutsche Bank analyst who, perhaps borrowing <a href="https://calnewport.com/the-dangers-of-vibe-reporting-about-ai/">​some of my​</a> terminology, <a href="https://www.nytimes.com/2026/02/25/business/citrini-ai-stock-market.html">​told the <em>Times</em>​</a> that the Citrini article had a “vibes-to-substance ratio” that was “undeniably high.”)</p>



<p>If you’re looking to reduce your blood pressure about this idea that AI is about to unravel the economy, I suggest reading <a href="https://www.citadelsecurities.com/news-and-insights/2026-global-intelligence-crisis/">​a detailed response article​</a> published by an analyst from the Global Macro Strategies group at Citadel. It begins with a bit of finance geek sarcasm:</p>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p>“Despite the macroeconomic community struggling to forecast 2-month-forward payroll growth with any reliable accuracy, the forward path of labor destruction can apparently be inferred with significant certainty from a hypothetical scenario posted on Substack…”</p>
</blockquote>



<p>It then continues to systematically destabilize the economic naivety of these breathless op-eds and viral essays about how AI will dismantle the economy all at once. It certainly made me feel better.</p>



<p>(If you’re looking for additional soothing of your AI anxiety, then you should also check <a href="https://www.youtube.com/watch?v=JRayjrpX10k">​the first episode​</a> of my new <em>AI Reality Check</em> podcast series, which I published last Thursday. I have a new episode of the series coming out this upcoming Thursday as well.)</p>
<p>The post <a href="https://calnewport.com/the-original-attention-crisis/">The Original Attention Crisis</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://calnewport.com/the-original-attention-crisis/feed/</wfw:commentRss>
			<slash:comments>4</slash:comments>
		
		
			</item>
		<item>
		<title>What Do Social Media Companies Fear? Time Management.</title>
		<link>https://calnewport.com/what-do-social-media-companies-fear-time-management/</link>
					<comments>https://calnewport.com/what-do-social-media-companies-fear-time-management/#comments</comments>
		
		<dc:creator><![CDATA[Study Hacks]]></dc:creator>
		<pubDate>Mon, 02 Mar 2026 11:00:00 +0000</pubDate>
				<category><![CDATA[Uncategorized]]></category>
		<guid isPermaLink="false">https://calnewport.com/?p=16806</guid>

					<description><![CDATA[<p>I recently came across an interesting academic article in the journal Frontiers in Psychology. It was titled, ​“The relationships between social media use, time management, ... <a title="What Do Social Media Companies Fear? Time Management." class="read-more" href="https://calnewport.com/what-do-social-media-companies-fear-time-management/" aria-label="Read more about What Do Social Media Companies Fear? Time Management.">Read more</a></p>
<p>The post <a href="https://calnewport.com/what-do-social-media-companies-fear-time-management/">What Do Social Media Companies Fear? Time Management.</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>I recently came across an interesting academic article in the journal <em>Frontiers in Psychology.</em> It was titled, <a href="https://www.frontiersin.org/journals/psychology/articles/10.3389/fpsyg.2025.1702767/full">​“The relationships between social media use, time management, and decision-making styles.”​</a></p>



<p>The paper’s author surveyed 612 university students and young adults, asking them, among other things, about their digital habits and levels of personal organization. Using a linear regression analysis, she uncovered the following:</p>



<p>“Social media use was negatively and significantly associated with overall time management and all its subscales.”</p>



<p>Here’s the standard interpretation of this result: Social media is distracting, and if you’re distracted, it becomes harder to maintain control over your schedule. So, the more you use social media, the worse you become at time management.</p>



<p>But I’ve become interested in the reverse form of this argument: <strong>the better your planning system, the less time you’ll spend on engagement-based applications like social media</strong>.</p>



<span id="more-16806"></span>



<p><em>Here’s my thinking…</em></p>



<p>When you’re following an intentional schedule, your efforts are oriented toward goals that you find important. You also feel a satisfying sense of self-efficacy. These realities engage your long-term reward system, which can override the urges generated by its short-term counterpart, dissipating the drive for quick gratification from activities like glancing at your phone.</p>



<p>In other words: The more you organize your analog life, the less appealing you’ll find the digital alternative.</p>



<p>If this is true, then maybe the thing social media companies fear most is not some newly-powerful application-blocking software or impossibly strict regulation, but rather a good old-fashioned daily planner.</p>



<figure class="wp-block-image"><img decoding="async" src="https://embed.filekitcdn.com/e/ekndSb6aixDTy6CAJEGkrv/pqwRyXLTHodBmNohEhe4Yo" alt=""/></figure>



<h3 class="wp-block-heading"><strong>In Other News</strong>:</h3>



<p>A lot of people I know have been freaked out recently by a viral essay with a grandiose title: <a href="https://x.com/mattshumer_/status/2021256989876109403">​“Something Big is Happening.”​</a> I recently released <a href="https://www.youtube.com/watch?v=Ijt8lV6b7QY">​a short video​</a> in which I conduct a close analysis of this piece. (Spoiler alert: I wasn’t impressed.) <a href="https://www.youtube.com/watch?v=Ijt8lV6b7QY">​<em>Check it out.</em>​</a></p>



<p>(More generally, I’ve been considering starting a separate weekly podcast/newsletter dedicated to providing a reality check on recent AI news. It feels like it might be useful to separate this discussion from my existing podcast and newsletter, which are more focused on how individuals can seek depth in a distracted world. But also, maybe this is a bad idea? I’m interested to hear your thoughts about this plan.)</p>
<p>The post <a href="https://calnewport.com/what-do-social-media-companies-fear-time-management/">What Do Social Media Companies Fear? Time Management.</a> appeared first on <a href="https://calnewport.com">Cal Newport</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://calnewport.com/what-do-social-media-companies-fear-time-management/feed/</wfw:commentRss>
			<slash:comments>19</slash:comments>
		
		
			</item>
	</channel>
</rss>

<!-- plugin=object-cache-pro client=phpredis metric#hits=2513 metric#misses=12 metric#hit-ratio=99.5 metric#bytes=802516 metric#prefetches=0 metric#store-reads=71 metric#store-writes=2 metric#store-hits=127 metric#store-misses=4 metric#sql-queries=27 metric#ms-total=339.71 metric#ms-cache=8.89 metric#ms-cache-avg=0.1235 metric#ms-cache-ratio=2.6 sample#redis-hits=407243619 sample#redis-misses=25586209 sample#redis-hit-ratio=94.1 sample#redis-ops-per-sec=32 sample#redis-evicted-keys=0 sample#redis-used-memory=68175104 sample#redis-used-memory-rss=59678720 sample#redis-memory-fragmentation-ratio=0.9 sample#redis-connected-clients=1 sample#redis-tracking-clients=0 sample#redis-rejected-connections=162 sample#redis-keys=81287 -->
