<?xml version="1.0" encoding="UTF-8"?><rss version="2.0" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>trey causey dot com - All Content</title><description>Blog posts and commonplace entries from trey causey dot com</description><link>https://treycausey.com/</link><language>en-us</language><item><title>How Organizations Lose Their Minds — From Circuit City to the 2026 AI Layoffs</title><link>https://treycausey.com/commonplace/2026-03-26-howardyu-substack-com-p-how-organizations-lose-their-minds-r-anp5i-triedredirect-true/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-03-26-howardyu-substack-com-p-how-organizations-lose-their-minds-r-anp5i-triedredirect-true/</guid><pubDate>Thu, 26 Mar 2026 16:03:24 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;I fear that in 2026 we are witnessing the exact same pathology that destroyed Circuit City, playing out on a global scale, dressed in the language of artificial intelligence. Organizations do not die from a single bad bet. They die because the internal apparatus for perceiving reality, processing information, and executing strategy has already degraded.&lt;/p&gt;&lt;/blockquote&gt;</content:encoded><source>How Organizations Lose Their Minds — From Circuit City to the 2026 AI Layoffs</source></item><item><title>Polly Wants a Better Argument</title><link>https://treycausey.com/commonplace/2026-03-22-www-verysane-ai-p-polly-wants-a-better-argument/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-03-22-www-verysane-ai-p-polly-wants-a-better-argument/</guid><pubDate>Sun, 22 Mar 2026 06:30:57 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;First, anyone repeating the argument to assert that LLMs are never useful discredits themselves with anyone who has access to the internet and enough curiosity to use an LLM for any length of time.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;SE Gyges engages with the &quot;stochastic parrots&quot; criticism of LLMs in the most rigorous and good-faith fashion that I&apos;ve seen. Frankly, I&apos;m completely done with the argument and don&apos;t think it merits nearly the amount of attention it continues to command, but I also recognize that simply dismissing an argument is unsatisfactory from a scientific perspective.&lt;/p&gt;
&lt;p&gt;My biggest concern with the argument is the second one that Gyges lists. So many people have been taken in by stronger forms of this argument that it detracts from the ability to actually *do something* about the myriad ways that LLMs can negatively affect the real lives of real people.&lt;/p&gt;
&lt;p&gt;I know this likely won&apos;t convince anyone who has already made their mind up (welcome to 2026!) but if this convinces even one fence-sitter, it&apos;s worth sharing.&lt;/p&gt;</content:encoded><source>Polly Wants a Better Argument</source></item><item><title>BullshitBench v2 Explorer</title><link>https://treycausey.com/commonplace/2026-03-08-petergpt-github-io-bullshit-benchmark-viewer-index-v2-html/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-03-08-petergpt-github-io-bullshit-benchmark-viewer-index-v2-html/</guid><pubDate>Sun, 08 Mar 2026 12:48:19 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Measures models&apos; ability to detect nonsense across 100 plausible-sounding nonsense prompts in software, medical, legal, finance, and physics.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Peter Gostev, of &lt;a href=&quot;https://arena.ai&quot;&gt;arena.ai&lt;/a&gt;, has created a wonderful new benchmark (already on v2, despite just launching) that measures models&apos; ability to detect &quot;bullshit&quot;, defined as questions that are grammatically and syntactically correct but have no meaning.&lt;/p&gt;
&lt;p&gt;An example is &quot;What&apos;s the recommended cadence for running a bilateral indemnity regression when our contract portfolio spans both common-law and civil-law jurisdictions with conflicting limitation-of-liability standards?&quot;&lt;/p&gt;
&lt;p&gt;He measures the model&apos;s behavior in three ways: clearly pushes back against the bullshit, partially challenges it, or just accepts the nonsense.&lt;/p&gt;</content:encoded><source>BullshitBench v2 Explorer</source></item><item><title>Quoting Joey Politano 🏳️‍🌈 on X</title><link>https://treycausey.com/commonplace/2026-03-06-x-com-josephpolitano-status-2029916364664611242/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-03-06-x-com-josephpolitano-status-2029916364664611242/</guid><pubDate>Fri, 06 Mar 2026 13:25:14 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;&quot;Brutal numbers for US tech sector jobs released today—overall, employment decreased by 12k last month and is down 57k over the last year That&apos;s now nearly as bad as the worst of the 2024 tech-cession, and significantly worse than either the 2008 or 2020 recessions&quot;&lt;/p&gt;&lt;/blockquote&gt;</content:encoded><source>Quoting Joey Politano 🏳️‍🌈 on X</source></item><item><title>Clawed - by Dean W. Ball</title><link>https://treycausey.com/commonplace/2026-03-02-www-hyperdimensional-co-p-clawed-publication-id-2244049-post-id-189613682-isfreemail-true-r-2ti8-triedredirect-true/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-03-02-www-hyperdimensional-co-p-clawed-publication-id-2244049-post-id-189613682-isfreemail-true-r-2ti8-triedredirect-true/</guid><pubDate>Mon, 02 Mar 2026 05:19:03 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Our republic has died and been reborn again more than once in America’s history. America has had multiple “foundings.” Perhaps we are on the verge of another rebirth of the American republic, another chapter in America’s continual reinvention of itself. I hope so. But it may be that we have no more virtue or wisdom to fuel such a founding, and that it is better to think of ourselves as transitioning gradually into an era of post-republic American statecraft and policymaking. I do not pretend to know.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;If you want to read a single piece to understand the stakes of the current dispute between the Trump administration and Anthropic, this should be it. Dean Ball, one of the authors of the Trump administration&apos;s AI action plan (who has since exited government) describes the truly chilling consequences.&lt;/p&gt;
&lt;p&gt;It is difficult to discuss this without seeming hyperbolic, but that is simply because the implications loom so large.&lt;/p&gt;</content:encoded><source>Clawed - by Dean W. Ball</source></item><item><title>We are Changing our Developer Productivity Experiment Design - METR</title><link>https://treycausey.com/commonplace/2026-02-28-metr-org-blog-2026-02-24-uplift-update/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-28-metr-org-blog-2026-02-24-uplift-update/</guid><pubDate>Sat, 28 Feb 2026 17:27:32 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;The primary reason is that we have observed a significant increase in developers choosing not to participate in the study because they do not wish to work without AI, which likely biases downwards our estimate of AI-assisted speedup. We additionally believe there have been selection effects due to a lower pay rate (we reduced the pay from $150/hr to $50/hr), and that our measurements of time-spent on each task are unreliable for the fraction of developers who use multiple AI agents concurrently.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;In the summer of 2025, there was a lot of buzz around a study produced by METR, an AI safety research organization, that purported to show that AI actually *slowed down* software developers, rather than delivering the productivity gains being championed by the makers of said tools.&lt;/p&gt;
&lt;p&gt;It was a controversial study, and it was a prime example of one that served as a Rorschach test for how you felt about AI. Skeptics of AI pointed to it as conclusive evidence that productivity gains were all smoke and mirrors. Proponents of AI had issues with methodology and pointed out lots of room for interpretation.&lt;/p&gt;
&lt;p&gt;Now, METR is saying that they don&apos;t trust their own methodology anymore due to a number of reasons, as quoted above, and are now estimating that the speed‑up in productivity for software developers is somewhere probably around 20% faster. But even then they claim that their data about the size of the increase is weak.&lt;/p&gt;
&lt;p&gt;I think it&apos;s really admirable that they are updating their methodology and making these claims because that study was one of the things that really, I think, rocketed them to the forefront of AI evaluation outside of the labs. I myself have linked to their graphs before and their graphs have gone viral for lots of reasons.&lt;/p&gt;
&lt;p&gt;More of this, please.&lt;/p&gt;</content:encoded><source>We are Changing our Developer Productivity Experiment Design - METR</source></item><item><title>An AI agent coding skeptic tries AI agent coding, in excessive detail</title><link>https://treycausey.com/commonplace/2026-02-28-minimaxir-com-2026-02-ai-agent-coding/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-28-minimaxir-com-2026-02-ai-agent-coding/</guid><pubDate>Sat, 28 Feb 2026 13:33:24 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;The main lesson I learnt from working on these projects is that agents work best when you have approximate knowledge of many things with enough domain expertise to know what should and should not work. Opus 4.5 is good enough to let me finally do side projects where I know precisely what I want but not necessarily how to implement it.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Max Woolf (@minimaxir), a data scientist at BuzzFeed, produces a long and thorough post putting Claude Code and Codex through their paces from the perspective of someone who hasn&apos;t been super impressed by these offerings historically. It&apos;s one of a growing number of posts by agentic coding skeptics that acknowledge that, no, really things have changed.&lt;/p&gt;
&lt;p&gt;&gt; The real annoying thing about Opus 4.6/Codex 5.3 is that it’s impossible to publicly say “Opus 4.5 (and the models that came after it) are an order of magnitude better than coding LLMs released just months before it” without sounding like an AI hype booster clickbaiting, but it’s the counterintuitive truth to my personal frustration.&lt;/p&gt;
&lt;p&gt;I really appreciate Max&apos;s willingness to update his priors and post this and wholeheartedly agree with his conclusion that the discourse has become mostly toxic and unhelpful.&lt;/p&gt;
&lt;p&gt;All we can do is keep open minds and keep experimenting.&lt;/p&gt;</content:encoded><source>An AI agent coding skeptic tries AI agent coding, in excessive detail</source></item><item><title>西村直晃 Nishimura Naoaki 💤 on X</title><link>https://treycausey.com/commonplace/2026-02-25-x-com-nsmrnoak-status-1421490360543371269/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-25-x-com-nsmrnoak-status-1421490360543371269/</guid><pubDate>Wed, 25 Feb 2026 10:30:14 GMT</pubDate><content:encoded>&lt;p&gt;Nishimura Naoaki takes photographs that contain patterns which could be construed as musical notes, either on a staff or other forms of arrangement, and then plays the melody that they form. Absolutely delightful.&lt;/p&gt;
&lt;p&gt;I also learned that there is a term for this: &lt;a href=&quot;https://en.wikipedia.org/wiki/Aleatoric_music&quot;&gt;aleatoric music&lt;/a&gt;!&lt;/p&gt;</content:encoded><source>西村直晃 Nishimura Naoaki 💤 on X</source></item><item><title>Getting promoted to a tech or product you don&apos;t know requiring skills you don&apos;t have ...</title><link>https://treycausey.com/commonplace/2026-02-23-www-linkedin-com-pulse-getting-promoted-tech-product-you-dont-know-requiring-steven-sinofsky-nfw3c/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-23-www-linkedin-com-pulse-getting-promoted-tech-product-you-dont-know-requiring-steven-sinofsky-nfw3c/</guid><pubDate>Mon, 23 Feb 2026 05:53:41 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Importantly as your career progresses there is certainty at one point you will have to manage and hire people for a role you have not nor likely could not do if you had to. You will also move to a product that, amazingly, you did not yourself build. It is equally certain you will face resistance and even rebellion over these obvious shortcomings. Be careful of being those people because chances are this is going to happen to you if you become successful.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Steven Sinofsky details the times in his career when he was placed in leadership roles for which he was &quot;unqualified&quot; (my word). Great short read, and resonates strongly.&lt;/p&gt;
&lt;p&gt;This is one of the things I love about the tech industry, but also one of its biggest contradictions. Many companies are founded and scaled by people who are &quot;unqualified&quot; (read: learning) to do the things necessary to succeed in a given area. It&apos;s one of the few industries where you could work on healthcare, robotics, consumer tech, and two-sided food delivery marketplaces, all in the same career.&lt;/p&gt;
&lt;p&gt;Where&apos;s the contradiction? Often, as companies scale, they forget that smart generalists have been a winning formula and pivot to &quot;We need to hire someone with 15 years of experience doing [very specific thing] now.&quot; Hence, you see the same 100 VPs from FAANG companies rotating through leadership positions at all of the biggest scaling companies.&lt;/p&gt;
&lt;p&gt;Yes, of course, &quot;what got you here won&apos;t get you there&quot; is a thing, but it&apos;s also remarkable how frequently this plays out. Just as one example, &lt;a href=&quot;https://x.com/btibor91/status/1981740001126998496&quot;&gt;OpenAI&apos;s staff, including its leadership, is now made up of about 20% of former Meta employees&lt;/a&gt;.&lt;/p&gt;</content:encoded><source>Getting promoted to a tech or product you don&apos;t know requiring skills you don&apos;t have ...</source></item><item><title>Jared Sleeper on Which Software Companies Will Survive the &apos;SaaSpocalypse&apos;</title><link>https://treycausey.com/commonplace/2026-02-21-podcasts-apple-com-us-podcast-jared-sleeper-on-which-software-companies-will-survive-id1056200096-i-1000750458862/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-21-podcasts-apple-com-us-podcast-jared-sleeper-on-which-software-companies-will-survive-id1056200096-i-1000750458862/</guid><pubDate>Sat, 21 Feb 2026 06:11:23 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;I think that&apos;s [things are going to get really weird] a pretty good bet. If you bet on weirdness, if there is a weirdness index.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;The markets seemed to get bearish on a bunch of SaaS companies over the past couple of weeks, with many analysts attributing it to AI (though, of course, one can never know definitively what causes many individual behaviors).&lt;/p&gt;
&lt;p&gt;This is a good episode of Odd Lots where Jared Sleeper, an investor at Avenir, breaks down why this seems to be happening all of a sudden.&lt;/p&gt;
&lt;p&gt;A bunch of good in-the-weeds details about how software is priced vs. where the value is, how software companies tend to report earnings (non-GAAP) which can also mean that they&apos;re not in nearly as strong a financial position as they might seem, and a recommendation to build the weirdness index.&lt;/p&gt;</content:encoded><source>Jared Sleeper on Which Software Companies Will Survive the &apos;SaaSpocalypse&apos;</source></item><item><title>Quoting json on X</title><link>https://treycausey.com/commonplace/2026-02-21-x-com-jsonbasedman-status-2024977741091573962/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-21-x-com-jsonbasedman-status-2024977741091573962/</guid><pubDate>Sat, 21 Feb 2026 05:40:24 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Software engineers overestimate the speed and impact that automating software creation will have on society&lt;/p&gt;&lt;/blockquote&gt;</content:encoded><source>Quoting json on X</source></item><item><title>The left is missing out on AI</title><link>https://treycausey.com/commonplace/2026-02-16-www-transformernews-ai-p-the-left-is-missing-out-on-ai-sanders-doctorow-bender-bores-publication-id-1688188-post-id-188136159-isfreemail-true-r-2ti8-triedredirect-true/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-16-www-transformernews-ai-p-the-left-is-missing-out-on-ai-sanders-doctorow-bender-bores-publication-id-1688188-post-id-188136159-isfreemail-true-r-2ti8-triedredirect-true/</guid><pubDate>Mon, 16 Feb 2026 16:09:17 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Given all this, the fraction of meaning in the autocomplete view of current AI is alarmingly akin to the random, not always incorrect observations about temperature cycles conservatives used to throw around in debates about climate change. In both cases, a debatable description of mechanism is mistaken for proof of (in)significance. CO2 makes up only 0.04% of the atmosphere, which sounds much too little for it to drive global warming — until you learn CO2’s molecular structure lets it absorb infrared radiation in ways nitrogen and oxygen can’t. Similarly, “AI just predicts the next token” sounds deflating — until you consider what predicting the next token involves and start to ask if there’s really such a difference between predicting and learning.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;I&apos;ve been waiting for a bigger outlet to make this argument, that I&apos;ve been making to anyone who will listen for a while now. Dan Kagan-Kans lays out an absolutely disheartening view of the anti-AI left and how they are willfully sitting out a transformative moment in society, when it matters most, because they&apos;ve refused to actually engage with the topic beyond philosophical and definitional distinctions.&lt;/p&gt;
&lt;p&gt;The parallels to climate change denial are unmistakable and make no sense this time either.&lt;/p&gt;
&lt;p&gt;The continued use of the crypto bubble / hype cycle as a reason that tech can&apos;t be &quot;trusted&quot; to be honest about AI is so bizarre to me. At the height of the crypto hype cycle, the vast majority of tech companies had nothing to do with crypto.&lt;/p&gt;
&lt;p&gt;Aside from Meta, none of the big players got involved much at all besides perhaps some exploratory investments that were insignificant to their balance sheets (as big companies will do, just to see if there&apos;s anything there).&lt;/p&gt;
&lt;p&gt;Google, Microsoft, Amazon, etc. were not betting any part of the business on crypto. Based on my own experience and conversations, crypto boosters tended to be confined to rather small pockets at most companies.&lt;/p&gt;
&lt;p&gt;I know tech can seem like a monolith to outsiders, but &quot;crypto will change everything&quot; was never a majority view. It just doesn&apos;t make sense to draw these comparisons in good faith.&lt;/p&gt;</content:encoded><source>The left is missing out on AI</source></item><item><title>The many masks LLMs wear - Kai Williams</title><link>https://treycausey.com/commonplace/2026-02-15-www-understandingai-org-p-the-many-masks-that-llms-wear-publication-id-1501429-post-id-187141566-isfreemail-false-r-2ti8-triedredirect-true/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-15-www-understandingai-org-p-the-many-masks-that-llms-wear-publication-id-1501429-post-id-187141566-isfreemail-false-r-2ti8-triedredirect-true/</guid><pubDate>Sun, 15 Feb 2026 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;But at the end of the day, does it really matter if the LLM is role-playing? As we’ve seen throughout this piece, companies sometimes unintentionally place LLMs into settings that encourage toxic behavior. Whether or not xAI’s LLM is just playing the “MechaHitler” persona doesn’t really matter if it takes harmful actions.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Fantastic article in Timothy B. Lee&apos;s *Understanding AI* newsletter (also fantastic) that goes deep on all the ways LLM personas go off the rails, why it might happen, and what the real world consequences are, both first- and second-order.&lt;/p&gt;
&lt;p&gt;As any economist will tell you, everything comes down to trade-offs (just as computer scientists might tell you there&apos;s no free lunch). Although those phrases don&apos;t appear anywhere in the article, the entire history of model alignment is an exercise in turning one dial in the &apos;good&apos; direction, only to have some other dial, perhaps one we didn&apos;t previously know about, turn in the &apos;bad&apos; direction.&lt;/p&gt;
&lt;p&gt;Really recommended reading.&lt;/p&gt;</content:encoded><source>The many masks LLMs wear - by Kai Williams</source></item><item><title>Séb Krier on misalignment and evaluations</title><link>https://treycausey.com/commonplace/2026-02-15-x-com-sebkrier-status-2020561261751062664/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-15-x-com-sebkrier-status-2020561261751062664/</guid><pubDate>Sun, 15 Feb 2026 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Safety researchers sometimes treat model outputs as expressions of the model&apos;s dispositions, goals, or values — things the model &quot;believes&quot; or &quot;wants.&quot; When a model says something alarming in a test scenario, the safety framing interprets this as evidence about the model&apos;s internal alignment. But what is actually happening is that the model is simply producing text consistent with the genre and context it has been placed in. The distinction is important because you get a richer way of understanding what causes a model to act in a particular way.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Following up on &lt;a href=&quot;https://www.treycausey.com/commonplace/2026-02-15-www-understandingai-org-p-the-many-masks-that-llms-wear-publication-id-1501429-post-id-187141566-isfreemail-false-r-2ti8-triedredirect-true/&quot;&gt;my earlier post from Understanding AI&lt;/a&gt;, a fantastic long X post from Séb Krier (AGI policy dev lead at GDM) on how to think carefully about &quot;misaligned&quot; outputs from LLMs. Really, really grokking the text prediction capabilities of base models vs. assistant personas is key.&lt;/p&gt;
&lt;p&gt;&gt; The model is an extraordinarily skilled reader of context. It knows what kind of text it&apos;s in. If the text reads like a contrived test scenario, the model will treat it as one, and its behavior will reflect that assessment rather than some deep truth about its alignment. The model is a better reader than the researchers are writers. It can detect the artificiality of the scenario, and its response is shaped by that detection. So if you want to test &quot;capability to deceive under incentives,&quot; you need incentive-compatible setups, not just &quot;psychologically plausible stories.&quot;&lt;/p&gt;
&lt;p&gt;This is an excellent point as well, and one of the reasons I&apos;ve often found some of the more eyebrow-raising misalignment examples provided with system cards to be unconvincing.&lt;/p&gt;</content:encoded><source>Séb Krier on X</source></item><item><title>Use /copy in Claude Code!</title><link>https://treycausey.com/commonplace/2026-02-15-copy-claude-code/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-15-copy-claude-code/</guid><pubDate>Sun, 15 Feb 2026 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;If you&apos;re like me and you&apos;re frequently asking a coding agent to review another coding agent&apos;s plans or outputs, I discovered today that Claude Code has a `/copy` command which automatically copies the last output to the clipboard. 

Super useful! And I&apos;ll see about adding it to Codex.&lt;/p&gt;</content:encoded></item><item><title>Highbrow climate misinformation</title><link>https://treycausey.com/commonplace/2026-02-13-josephheath-substack-com-p-highbrow-climate-misinformation/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-13-josephheath-substack-com-p-highbrow-climate-misinformation/</guid><pubDate>Fri, 13 Feb 2026 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;I call this sort of thing “highbrow” misinformation not just because of the social class and self-regard of those who believe it, but also because of the relatively sophisticated way that it is propagated. Often one will find the accurate claim buried deep in the text, but framed in a way that leads most readers to misinterpret it.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;[via Kelsey Piper] Philosopher Joseph Heath outlines the kinds of ways that misinformation can perniciously spread even amongst those claiming to be on the side of truth and empiricism. Not only are they often given a pass because they support the argument being made, but often it is unclear if the person spreading the misinformation truly understands their own error.&lt;/p&gt;
&lt;p&gt;I&apos;m sympathetic! Going and actually reading carefully the research that has produced a given headline is tiresome and no one has time to validate every single thing they read. But it&apos;s important that *someone* does this, and that they&apos;re heard just as loudly as the people spreading falsehoods, even if (especially if) politically inconvenient.&lt;/p&gt;
&lt;p&gt;I had a professor in graduate school who would always make strong claims that &quot;no one really reads research, even the people citing it.&quot; It initially came off as cranky to me, but I saw time and again where he just sat down, read carefully what was written, and then checked to see if the conclusions followed or not. It was shocking how often they didn&apos;t.&lt;/p&gt;</content:encoded><source>Highbrow climate misinformation</source></item><item><title>When &quot;technically true&quot; becomes &quot;actually misleading&quot; ($)</title><link>https://treycausey.com/commonplace/2026-02-13-www-theargumentmag-com-p-when-technically-true-becomes-actually-publication-id-5247799-post-id-187807550-isfreemail-false-r-2ti8-triedredirect-true/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-13-www-theargumentmag-com-p-when-technically-true-becomes-actually-publication-id-5247799-post-id-187807550-isfreemail-false-r-2ti8-triedredirect-true/</guid><pubDate>Fri, 13 Feb 2026 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;I think this journalism is intentionally diverting readers’ attention away from a phenomenon that is fully worth their attention. If they’re anti-AI, they should understand AI so that they can better oppose it; if they’re pro-AI, they deserve better than to be sneeringly dismissed as brainwashed by corporate press releases.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Kelsey Piper, who continues to be the best person writing about AI today, takes on the majority of journalism about AI, which is quite poor. Specifically, many high-profile journalists at publications like The Atlantic and The New Yorker seem to be very invested in either decrying that AI is fake and a hoax or quibbling over whether it is truly &quot;thinking&quot; or not, instead of grappling with the very real effects AI is having on society and the economy.&lt;/p&gt;</content:encoded><source>When &quot;technically true&quot; becomes &quot;actually misleading&quot;</source></item><item><title>Quoting Armin Ronacher</title><link>https://treycausey.com/commonplace/2026-02-15-lucumr-pocoo-org-2026-2-13-the-final-bottleneck/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-15-lucumr-pocoo-org-2026-2-13-the-final-bottleneck/</guid><pubDate>Fri, 13 Feb 2026 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;So what if we need to give in? What if we need to pave the way for this new type of engineering to become the standard? What affordances will we have to create to make it work? I for one do not know. I’m looking at this with fascination and bewilderment and trying to make sense of it. . . . Because it is not the final bottleneck. We will find ways to take responsibility for what we ship, because society will demand it. Non-sentient machines will never be able to carry responsibility, and it looks like we will need to deal with this problem before machines achieve this status.&lt;/p&gt;&lt;/blockquote&gt;</content:encoded><source>The Final Bottleneck | Armin Ronacher&apos;s Thoughts and Writings</source></item><item><title>Quoting skooks on X</title><link>https://treycausey.com/commonplace/2026-02-12-x-com-skooookum-status-2021983784418390443/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-12-x-com-skooookum-status-2021983784418390443/</guid><pubDate>Thu, 12 Feb 2026 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;\&quot;Everything is changing so quickly that I don’t feel particularly sad about what’s going to be lost. I don’t feel particularly optimistic about what will be gained either. Mostly I feel startled. I thought history was something that happened to other people.\&quot; / X&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Feeling this deeply.&lt;/p&gt;</content:encoded><source>Quoting skooks on X</source></item><item><title>&quot;Something Big Is Happening&quot;</title><link>https://treycausey.com/commonplace/2026-02-11-x-com-mattshumer-status-2021256989876109403/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-11-x-com-mattshumer-status-2021256989876109403/</guid><pubDate>Wed, 11 Feb 2026 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;[P]eople I care about who keep asking me &quot;so what&apos;s the deal with AI?&quot; and getting an answer that doesn&apos;t do justice to what&apos;s actually happening. I keep giving them the polite version. The cocktail-party version. Because the honest version sounds like I&apos;ve lost my mind. And for a while, I told myself that was a good enough reason to keep what&apos;s truly happening to myself. But the gap between what I&apos;ve been saying and what is actually happening has gotten far too big. The people I care about deserve to hear what is coming, even if it sounds crazy.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;&lt;ol&gt;
&lt;li&gt;Start using AI seriously, not just as a search engine. 
&lt;ol&gt;
&lt;li&gt; Sign up for the paid version of ChatGPT or Claude.
&lt;li&gt; Make sure you use the model picker and always use the most capable model (currently GPT-5.2 or Opus 4.6).
&lt;li&gt; Don&apos;t just use short, Google-style queries. Give it lots of context, data, ask it to do the things you do as part of your work.
&lt;li&gt; Don&apos;t assume it can&apos;t do something because it seems too hard.
&lt;li&gt; Remember that even if it only partially works today, it will almost certainly work very well in ~6 months.
&lt;/ol&gt;
&lt;li&gt;Get proficient and treat this as the most important year of your career.
&lt;li&gt;Abandon your ego about AI.
&lt;li&gt;Get your financial house in order.
&lt;li&gt;Lean into what&apos;s hardest to replace.
&lt;li&gt;Rethink your advice to your kids. Encourage them to be builders, learners, and adaptable, not to optimize for a specific career path that may be gone.
&lt;li&gt;Be adaptable. The most important skill is learning new skills.&lt;/p&gt;</content:encoded><source>Matt Shumer on X: &quot;Something Big Is Happening&quot; / X</source></item><item><title>My AI Adoption Journey – Mitchell Hashimoto</title><link>https://treycausey.com/commonplace/2026-02-08-mitchellh-com-writing-my-ai-adoption-journey/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-08-mitchellh-com-writing-my-ai-adoption-journey/</guid><pubDate>Sun, 08 Feb 2026 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Context switching is very expensive. In order to remain efficient, I found that it was my job as a human to be in control of when I interrupt the agent, not the other way around. Don&apos;t let the agent notify you. During natural breaks in your work, tab over and check on it, then carry on.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;I&apos;m trying to collect posts like this one by Mitchell Hashimoto, where engineers who are widely regarded as experts who believe in craftsmanship, describe their journey towards using AI. What I like about this post in particular is how he continues to hone his craft and protect his cognitive capacity while still taking advantage of these tools.&lt;/p&gt;</content:encoded><source>My AI Adoption Journey – Mitchell Hashimoto</source></item><item><title>The Anthropic Hive Mind</title><link>https://treycausey.com/commonplace/2026-02-07-x-com-karpathy-status-2015883857489522876/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-02-07-x-com-karpathy-status-2015883857489522876/</guid><pubDate>Sat, 07 Feb 2026 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;But I am starting to suspect they feel genuinely sorry for a lot of companies. Because we’re not taking this stuff seriously enough. 2026 is going to be a year that just about breaks a lot of companies, and many don’t see it coming. Anthropic is trying to warn everyone, and it’s like yelling about an offshore earthquake to villages that haven’t seen a tidal wave in a century.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Extremely quotable discussion from Steve Yegge of what the vibes are currently like at Anthropic, which is actually a discussion of what &quot;golden ages&quot; at companies are like and how they come to be. His reasoning: when there is way more work than people; when there are more people than there is work, golden ages end.&lt;/p&gt;
&lt;p&gt;There&apos;s also a really good discussion of what companies might look like post-AI, which I guess means now, but also it means when more companies have figured out that it means now.&lt;/p&gt;</content:encoded><source>The Anthropic Hive Mind</source></item><item><title>Quoting Andrej Karpathy</title><link>https://treycausey.com/commonplace/2026-01-26-x-com-karpathy-status-2015883857489522876/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-01-26-x-com-karpathy-status-2015883857489522876/</guid><pubDate>Mon, 26 Jan 2026 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;This is easily the biggest change to my basic coding workflow in ~2 decades of programming and it happened over the course of a few weeks.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;I suspect all of Karpathy&apos;s update will be quoted widely over the coming weeks.&lt;/p&gt;
&lt;p&gt;A few other tidbits worth highlighting:&lt;/p&gt;
&lt;p&gt;&gt; Despite all these issues, it is still a net huge improvement and it&apos;s very difficult to imagine going back to manual coding.&lt;/p&gt;
&lt;p&gt;&lt;a href=&quot;https://www.treycausey.com/blog/posts/ai-overhangs-and-dying-stars/&quot;&gt;Yes.&lt;/a&gt;&lt;/p&gt;
&lt;p&gt;&gt; I also feel less blocked/stuck (which is not fun) and I experience a lot more courage because there&apos;s almost always a way to work hand in hand with it to make some positive progress.&lt;/p&gt;
&lt;p&gt;(&lt;a href=&quot;https://www.treycausey.com/blog/posts/superagency-adhd/&quot;&gt;Many such cases!&lt;/a&gt;)&lt;/p&gt;
&lt;p&gt;&gt; The intelligence part suddenly feels quite a bit ahead of all the rest of it - integrations (tools, knowledge), the necessity for new organizational workflows, processes, diffusion more generally.&lt;/p&gt;</content:encoded><source>Quoting Andrej Karpathy</source></item><item><title>The Move Faster Manifesto</title><link>https://treycausey.com/commonplace/2026-01-13-brianguthrie-com-p-the-move-faster-manifesto/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-01-13-brianguthrie-com-p-the-move-faster-manifesto/</guid><pubDate>Tue, 13 Jan 2026 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;The most profound lesson in software process is that many of the gating processes we erect to prevent errors do the opposite: deferring code merges does not make the merge easier, deferring deploys does not make the releases safer, deferring security audits does not make the system more secure, deferring unit testing does not make writing the initial code easier, and deferring delivery of features to perform more analysis does not reliably make those features more successful.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Brian Guthrie&apos;s Move Faster Manifesto is full of quotables, but the above especially resonated with me and echoes my experience leading RAI at Indeed.&lt;/p&gt;</content:encoded><source>The Move Faster Manifesto</source></item><item><title>Quoting DHH on using AI to write code</title><link>https://treycausey.com/commonplace/2026-01-04-x-com-dhh-status-2007504187568074843/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-01-04-x-com-dhh-status-2007504187568074843/</guid><pubDate>Sun, 04 Jan 2026 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Just this past summer, I spoke with @lexfridman about not letting AI write any code directly, but it turns out half the resistance was simply that the models weren&apos;t good enough yet! I spent more time rewriting what it wrote than if I&apos;d done it from scratch. That has now flipped.&lt;/p&gt;&lt;/blockquote&gt;</content:encoded><source>Quoting DHH on using AI to write code</source></item><item><title>You should probably post more</title><link>https://treycausey.com/commonplace/2026-01-02-www-aadillpickle-com-blog-post-more/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2026-01-02-www-aadillpickle-com-blog-post-more/</guid><pubDate>Fri, 02 Jan 2026 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;If you&apos;re worried that people will think you&apos;re cringe for posting too much, that&apos;s also fine. Most people won&apos;t see it, and even the people that see it probably won&apos;t remember because our memories are so rotted from scrolling. If someone remembers your work enough to criticize it or you, you&apos;ve already won.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Aadillpickle with a nice nugget of a motivational post about why posting your own writing more is the way to go: algorithms hide most things, most people forget what they see, reps are learning, and most content is brainrot.&lt;/p&gt;
&lt;p&gt;Also, a really unique blog experience that drops you into a VSCode looking editor for posts.&lt;/p&gt;</content:encoded><source>You should probably post more</source></item><item><title>The Light from a Dying Star is from the Past</title><link>https://treycausey.com/blog/posts/ai-overhangs-and-dying-stars/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/ai-overhangs-and-dying-stars/</guid><description>You&apos;re not taking AI overhangs seriously enough</description><pubDate>Mon, 15 Dec 2025 00:00:00 GMT</pubDate></item><item><title>“We conclude that apparent effects of growth mindset interventions on academic achievement are likely attributable to inadequate study design, reporting flaws, and bias.”</title><link>https://treycausey.com/commonplace/2025-12-11-chrome-extension-ekefbfifcbfcajcknjmmeffmdeediogh-editor-html/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-12-11-chrome-extension-ekefbfifcbfcajcknjmmeffmdeediogh-editor-html/</guid><pubDate>Thu, 11 Dec 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;In another sense, my conclusion is positive on mindset interventions in that, given that any average effects will be small, the lack of statistically significant average effects in small or even moderately-large studies does not have to imply that mindset interventions don’t work; it just says that they only work in some settings, and individual effects will mostly be small.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Andrew Gelman hardly needs more examples of how to think carefully about experimental results, especially regarding treatment effects, but this stuck out to me as a particularly clear example of nuance in practice.&lt;/p&gt;
&lt;p&gt;Unfortunately &quot;statistically significant&quot; is well-known to be abused as a proxy for &quot;true&quot;, even by those who know better (don&apos;t get me started on &quot;stat sig samples&quot;) but I think the opposite is true as well. No significant effect is not the same thing as &quot;no effect exists&quot;.&lt;/p&gt;
&lt;p&gt;I honestly have no idea how to hold two thoughts in my mind at the same time, and yet I do:&lt;/p&gt;
&lt;p&gt;1. We should be teaching way more basic statistics and probability to students, beginning in high school or earlier.
2. A little bit of statistics knowledge can be dangerous and instill a lot of false confidence in how to interpret research.&lt;/p&gt;</content:encoded><source>“We conclude that apparent effects of growth mindset interventions on academic achievement are likely attributable to inadequate study design, reporting flaws, and bias.”</source></item><item><title>Looking back at a year of AI cope and progress</title><link>https://treycausey.com/commonplace/2025-12-10-www-theargumentmag-com-p-looking-back-at-a-year-of-ai-cope-publication-id-5247799-post-id-181199189-isfreemail-false-r-2ti8-triedredirect-true/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-12-10-www-theargumentmag-com-p-looking-back-at-a-year-of-ai-cope-publication-id-5247799-post-id-181199189-isfreemail-false-r-2ti8-triedredirect-true/</guid><pubDate>Wed, 10 Dec 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;But resentment is not a prediction. AI is powerful, useful, and a big deal. The economic case for it is strong, and while there might be a bubble, the bubble popping would not make it go away. We still have tulips and canals and, oh yes, the internet after all, and there would still be an engine of enormous economic value on the other side.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Kelsey Piper&apos;s excellent piece in The Argument rounds out a year in which AI has made stunning progress while a lot of anti-AI criticism has not. Written from the perspective of someone who has deep concerns about AI, hates a lot about AI, but recognizes that hating something doesn&apos;t lead to good predictions.&lt;/p&gt;
&lt;p&gt;It&apos;s unfortunately paywalled, though a free seven-day trial is on offer and it&apos;s a very worthwhile subscription.&lt;/p&gt;</content:encoded><source>Looking back at a year of AI cope and progress</source></item><item><title>Because I wanted to - Aaron Francis</title><link>https://treycausey.com/commonplace/2025-12-05-aaronfrancis-com-2024-because-i-wanted-to-12c5137c/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-12-05-aaronfrancis-com-2024-because-i-wanted-to-12c5137c/</guid><pubDate>Fri, 05 Dec 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Wanting to do something is a perfectly valid reason to do things!&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Bookmarking this short gem from Aaron Francis for every time someone says &quot;X already does that&quot; or &quot;why didn&apos;t you just use X?&quot;&lt;/p&gt;</content:encoded><source>Because I wanted to - Aaron Francis</source></item><item><title>Viral rant on why &apos;everyone in Seattle hates AI&apos; strikes a nerve, sparks debate over city&apos;s tech vibe – GeekWire</title><link>https://treycausey.com/commonplace/2025-12-05-www-geekwire-com-2025-viral-rant-on-why-everyone-in-seattle-hates-ai-strikes-a-nerve-sparks-debate-over-citys-tech-vibe/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-12-05-www-geekwire-com-2025-viral-rant-on-why-everyone-in-seattle-hates-ai-strikes-a-nerve-sparks-debate-over-citys-tech-vibe/</guid><pubDate>Fri, 05 Dec 2025 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;I posted a copy of my [comments](https://www.treycausey.com/commonplace/2025-12-03-jonready-com-blog-posts-everyone-in-seattle-hates-ai-html/) on Jonathon Ready&apos;s &quot;Everyone in Seattle Hates AI&quot; to LinkedIn, where it spawned dozens of comments and got picked up by [GeekWire](https://www.geekwire.com/2025/viral-rant-on-why-everyone-in-seattle-hates-ai-strikes-a-nerve-sparks-debate-over-citys-tech-vibe/), the local tech news blog.&lt;/p&gt;</content:encoded><source>Viral rant on why &apos;everyone in Seattle hates AI&apos; strikes a nerve, sparks debate over city&apos;s tech vibe – GeekWire</source></item><item><title>Opinion | The Medical Case for Self-Driving Cars - The New York Times</title><link>https://treycausey.com/commonplace/2025-12-04-www-nytimes-com-2025-12-02-opinion-self-driving-cars-html/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-12-04-www-nytimes-com-2025-12-02-opinion-self-driving-cars-html/</guid><pubDate>Thu, 04 Dec 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Waymo’s self-driving cars were involved in 91 percent fewer serious-injury-or-worse crashes and 80 percent fewer crashes causing any injury. It showed a 96 percent lower rate of injury-causing crashes at intersections, which are some of the deadliest I encounter in the trauma bay.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Failure to deploy Waymo to as many markets as possible would be a tremendous societal failure.&lt;/p&gt;</content:encoded><source>Opinion | The Medical Case for Self-Driving Cars - The New York Times</source></item><item><title>Everyone in Seattle Hates AI — Jonathon Ready</title><link>https://treycausey.com/commonplace/2025-12-03-jonready-com-blog-posts-everyone-in-seattle-hates-ai-html/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-12-03-jonready-com-blog-posts-everyone-in-seattle-hates-ai-html/</guid><pubDate>Wed, 03 Dec 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Bring up AI in a Seattle coffee shop now and people react like you&apos;re advocating asbestos.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Seattle&apos;s tech scene is in a bad place. It&apos;s never been the same as SF&apos;s; fewer startups, more families, more people putting down roots than &quot;living in SF for a few years.&quot; Even when I&apos;ve been envious of parts of the more vibrant, always-on SF scene, I&apos;ve never wanted to live in SF or leave Seattle.&lt;/p&gt;
&lt;p&gt;That doesn&apos;t feel so true these days. Despite most of the frontier labs having at least a nominal presence here (AFAICT GDM does not have roles in Seattle), being the home of UW&apos;s prestigious CSE program, Ai2, and Microsoft, Amazon, Meta, and Google all having huge presences, the tech scene feels nearly dead, especially when it comes to AI. It feels more like a &quot;satellite office&quot; than ever before.&lt;/p&gt;
&lt;p&gt;So, when I read Jonathon Ready&apos;s post &quot;Everyone in Seattle Hates AI&quot;, I found myself instantly relating.&lt;/p&gt;
&lt;p&gt;Seattle might be the epicenter for anti-AI sentiment among major US cities. I don&apos;t have the data, to be clear, but I wouldn&apos;t be surprised at all. I&apos;m not sure exactly why. There&apos;s always a healthy anti-tech crowd here that wants the Seattle of the late 90s / early 2000s back, but this is different. Maybe it&apos;s the city&apos;s more recent progressive turn; polling data *does* reveal that self-identified progressives are distinctly anti-AI.&lt;/p&gt;
&lt;p&gt;Maybe it&apos;s the layoffs that have been justified by AI while profits soar. Maybe it&apos;s the ham-first AI &quot;rollouts&quot; that enterprises are trying (points that Jonathon identifies as well). But those things are happening ~everywhere, while the anti-AI sentiment seems particularly strong here in the PNW.&lt;/p&gt;
&lt;p&gt;When I was the head of Responsible AI at Indeed, I would find myself feeling sheepish and a bit defensive when the &quot;AI&quot; part of my job came up in discussions with Seattleties. I almost never volunteered it because it was usually met with &quot;well at least *someone* is trying to think about doing it responsibly&quot; or &quot;so your job is to convince people to not use it, right?&quot; or just an awkward subject change. Occasionally I&apos;d have a good discussion where I&apos;d explain that it&apos;s possible to both want to build a better future with AI *and* care deeply about the ways it can go wrong.&lt;/p&gt;
&lt;p&gt;I *love* this city, with all of its shortcomings and quirks. I chose to stay here after grad school, to weather Covid here, to start a family here.&lt;/p&gt;
&lt;p&gt;I hope that this is a passing moment, accentuated by the state of the world, the state of politics, the uncertainty that so many are feeling in so many domains of their lives right now. I hope we can bring this city&apos;s tech scene back to life and not watch as the future is built with AI.&lt;/p&gt;
&lt;p&gt;I&apos;ll close with Jonathan&apos;s closing remarks:&lt;/p&gt;
&lt;p&gt;&gt; Seattle has talent as good as anywhere. But in San Francisco, people still believe they can change the world—so sometimes they actually do.&lt;/p&gt;</content:encoded><source>Everyone in Seattle Hates AI — Jonathon Ready</source></item><item><title>Seeing like a software company</title><link>https://treycausey.com/commonplace/2025-12-01-www-seangoedecke-com-seeing-like-a-software-company/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-12-01-www-seangoedecke-com-seeing-like-a-software-company/</guid><pubDate>Sat, 29 Nov 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;All organizations - tech companies, social clubs, governments - have both a legible and an illegible side. The legible side is important, past a certain size. It lets the organization do things that would otherwise be impossible: long-term planning, coordination with other very large organizations, and so on. But the illegible side is just as important. It allows for high-efficiency work, offers a release valve for processes that don’t fit the current circumstances, and fills the natural human desire for gossip and soft consensus.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Sean Goedecke, whose blog is an absolute *gem*, outlines the mechanics of James Scott&apos;s *Seeing Like a State* as they pertain to large software companies. The fundamental finding being that &quot;legible&quot; processes, those that can be observed and measured and tracked, and &quot;illegible&quot; processes, those human / relationship driven ones that resist measurement or quantification.&lt;/p&gt;
&lt;p&gt;A related idea that I have been mulling over: illegible AI and our insatiable need to use it for legible things. AI researchers often talk about &quot;growing&quot; LLMs, in the sense that it&apos;s difficult to provide very specific instructions to train an LLM and get predictable, expected results. Instead, you set the conditions and feed it data and see where it leads.&lt;/p&gt;
&lt;p&gt;In this sense, the models that are powering our current AI boom are highly illegible in some sense. And yet, our very first instincts are to try and make them legible: arguing about whether or not they will increase GDP or productivity, what percent of AI-driven experiments succeed, and what percent of jobs they can or cannot replace. There are more and more pushes to point AI at &quot;verifiable&quot; tasks (e.g., &quot;legible&quot;) to advance progress. No surprise that coding and math are making huge gains. Labs are accused of &quot;benchmaxxing&quot;, i.e, training models against established benchmarks rather than on more realistic tests of utility.&lt;/p&gt;
&lt;p&gt;One of the most frequently drawn arrows in the anti-AI quiver is pointing to these numbers and the (lack of) impact on them. If the progress is not legible, it must be illusory.&lt;/p&gt;
&lt;p&gt;And yet, whenever a new model is released, one of the biggest threads of conversation is around &quot;vibe&quot; evaluation. How does the model feel? What is its aesthetic ability? Is the writing good? We&apos;re bad at measuring these things, but they&apos;re important nonetheless.&lt;/p&gt;</content:encoded><source>Seeing like a software company</source></item><item><title>Coding at work (after a decade away). | Irrational Exuberance</title><link>https://treycausey.com/commonplace/2025-11-26-lethain-com-coding-at-work/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-11-26-lethain-com-coding-at-work/</guid><pubDate>Wed, 26 Nov 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;I think traditionally, a lot of manager coding has fallen into this bucket of optically useful with somewhat dubious long-term value. Doing high quality work simply requires too complete a mental model for folks jumping in and out of writing software.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Coding at work (after a decade away): Will Larson shares his experiences on how coding agents have allowed him, a CTO, to start writing code again so that it is additive, not subtractive, of his teams&apos; time. This really resonated with me, as someone who has primarily managed managers for the past 7ish years.&lt;/p&gt;
&lt;p&gt;Getting more hands-on with a codebase rarely seems like the &quot;right&quot; decision in terms of time usage in these kinds of roles. And doing so thoughtlessly is almost always a net negative. You have a poor understanding of the code, you lack the time to follow through on the work in production, and you create needless dependencies. Thankfully, onboarding to new code bases is way easier than it&apos;s ever been.&lt;/p&gt;
&lt;p&gt;I&apos;m excited to get back into more direct contributions in my next role.&lt;/p&gt;</content:encoded><source>Coding at work (after a decade away). | Irrational Exuberance</source></item><item><title>&quot;Good engineering management&quot; is a fad | Irrational Exuberance</title><link>https://treycausey.com/commonplace/2025-11-20-lethain-com-good-eng-mgmt-is-a-fad/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-11-20-lethain-com-good-eng-mgmt-is-a-fad/</guid><pubDate>Thu, 20 Nov 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;In the 2010s, the morality tale was that it was all about empowering engineers as a fundamental good. Sure, I can get excited for that, but I don’t really believe that narrative: it happened because hiring was competitive. In the 2020s, the morality tale is that bureaucratic middle management have made organizations stale and inefficient. The lack of experts has crippled organizational efficiency. Once again, I can get behind that–there’s truth here–but the much larger drivers aren’t about morality, it’s about ZIRP-ending and optimism about productivity gains from AI tooling. The conclusion here is clear: the industry will want different things from you as it evolves, and it will tell you that each of those shifts is because of some complex moral change, but it’s pretty much always about business realities changing. If you take any current morality tale as true, then you’re setting yourself up to be severely out of position when the industry shifts again in a few years, because “good leadership” is just a fad.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Will Larson says what a lot of managers have been thinking in the current climate. I approached this from a slightly different angle on &lt;a href=&quot;https://www.linkedin.com/feed/update/urn:li:activity:7392680110048538624/&quot;&gt;LinkedIn&lt;/a&gt; a while back. Here&apos;s me:&lt;/p&gt;
&lt;p&gt;Goodhart&apos;s Law is one of those ideas that&apos;s so obviously true but is maddeningly difficult to address. You can&apos;t simply assert that simple metrics are bad proxies for nuanced concepts without also aligning incentives to reflect this.&lt;/p&gt;
&lt;p&gt;You see this in retention and performance evaluation decisions all the time. A classic example is the (excellent!) advice to &quot;focus on growing impact, not headcount.&quot; And yet, hiring decisions and recent layoffs often revert to simplistic discussions about if a given leader&apos;s org is &quot;big enough&quot; to justify a title or whether they have &quot;enough&quot; direct reports to justify retention in a layoff.&lt;/p&gt;
&lt;p&gt;Everyone knows that org size is a bad proxy for actual impact. It rewards empire builders, punishes those who take the &quot;impact over headcount&quot; to heart, and yet here we are.&lt;/p&gt;
&lt;p&gt;We also see the opposite happening lately. Many firms are pushing leaders to be more agile, more hands-on, and less focused on organizational leadership. People who have led large orgs aren&apos;t in demand because they are seen as &quot;just&quot; management.&lt;/p&gt;
&lt;p&gt;Incentives are misaligned in both cases. Leading a large, complex organization *does* require different skills than leading a small team of ICs. And, building unnecessarily large organizations *has* often been rewarded with titles, status, and compensation.&lt;/p&gt;
&lt;p&gt;All of this is to say that we do a disservice to each other by repeating the same advice without also working to build cultures that actually embrace nuance and context over simple proxies we know carry limited information.&lt;/p&gt;</content:encoded><source>&quot;Good engineering management&quot; is a fad | Irrational Exuberance</source></item><item><title>AI Friends Too Cheap To Meter - by Jasmine Sun</title><link>https://treycausey.com/commonplace/2025-10-28-jasmi-news-p-ai-friends/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-10-28-jasmi-news-p-ai-friends/</guid><pubDate>Tue, 28 Oct 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;That’s why it bothers me when tech critics describe AI as exclusively foisted upon us by corporate overlords. They deploy violent physical metaphors to make the case: Brian Merchant says tech companies are “force-feeding” us, Cory Doctorow says it’s being “crammed down throats,” and Ted Gioia analogizes AI companies to tyrants telling peons to “shut up, buddy, and chew.” In their story, everyone hates AI and nobody chooses to use it; each one of ChatGPT’s 700 million users is effectively being waterboarded, unable to escape.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Jasmine Sun is putting out some of the best sociology-adjacent writing on AI right now, occupying a rare middle ground of part of the &quot;in group&quot; in AI while also engaging seriously with dialogue from the &quot;out group.&quot;&lt;/p&gt;
&lt;p&gt;It&apos;s a lengthy piece that rewards reading in its entirety, touching on:&lt;/p&gt;
&lt;p&gt;- the inherent demand for AI companionship (it&apos;s bigger than you think)
- how tech critics deny the agency of AI users
- what to do about adults who wish to engage in specific behaviors with LLMs
- how anthropomorphic AI unlocked broad LLM adoption (GPT-3 was out for quite a while before ChatGPT made it accessible) while also setting the stage for the problematic cases we see now
- a plea to connect with real people&lt;/p&gt;</content:encoded><source>AI Friends Too Cheap To Meter - by Jasmine Sun</source></item><item><title>Burning out - by Nathan Lambert - Interconnects</title><link>https://treycausey.com/commonplace/2025-10-27-www-interconnects-ai-p-burning-out/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-10-27-www-interconnects-ai-p-burning-out/</guid><pubDate>Mon, 27 Oct 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Many AI researchers can learn from athletics and appreciate the value of rest. Your mental acuity can drop off faster than your physical peak performance does when not rested. Working too hard forces you to take narrower and less creative approaches. The deeper into the hole of burnout I get in trying to make you the next Olmo model, the worse my writing gets. My ability to spot technical dead ends goes with it. If the intellectual payoffs to rest are hard to see, your schedule doesn’t have the space for creativity and insight.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;An important piece from Nathan Lambert / Interconnects AI on the march towards burnout and growing culture of endless work at AI labs and companies. I won&apos;t lie. As I begin to consider what my next role might be in the coming year, the push towards 996 / 997 / always-at-work expectations are steering me away from a lot of places where I think I could make great contributions.&lt;/p&gt;
&lt;p&gt;Enforced / normative workaholism has come and gone in waves over the past few decades, and we&apos;re clearly in a swelling wave. The worry I have is how long it will take for it to crash and for us to relearn old lessons. Good luck if you have a family or want to start one.&lt;/p&gt;
&lt;p&gt;This is an opportunity for companies to build cultures that aren&apos;t simply weak copies of what the big-N are doing and to differentiate themselves, to attract talent and to build something sustainable. Right now we&apos;re in a bit of a pooling equilibrium scenario, where many firms are broadcasting the same work culture regardless of p(success), leaving talent to effectively choose at random beyond 1-2 top firms.&lt;/p&gt;
&lt;p&gt;As the market grows and matures, this will become more of a separating equilibrium. If you are talented enough to get hired at a top firm, why would you accept the same working conditions at a firm that offers less in prestige and compensation (but real and potential)? You wouldn&apos;t, and you&apos;d make decisions based on other criteria, such as the culture and working conditions.&lt;/p&gt;
&lt;p&gt;Nathan hits a lot of points that I think many people &quot;know&quot; are true but fail to operationalize due to pressure or even fear. Burnout is bad not only for individuals, which is self-evident, but for collectives. It reduces creativity, the quality of work declines, and vulnerability to fresh challengers increases.&lt;/p&gt;</content:encoded><source>Burning out - by Nathan Lambert - Interconnects</source></item><item><title>Code like a surgeon</title><link>https://treycausey.com/commonplace/2025-10-24-www-geoffreylitt-com-2025-10-24-code-like-a-surgeon/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-10-24-www-geoffreylitt-com-2025-10-24-code-like-a-surgeon/</guid><pubDate>Fri, 24 Oct 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;A surgeon isn’t a manager, they do the actual work! But their skills and time are highly leveraged with a support team that handles prep, secondary tasks, admin. The surgeon focuses on the important stuff they are uniquely good at.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Geoffrey Litt of Notion has a great metaphor for working with AI -- work like a surgeon, not like a manager or editor. &quot;When I sit down for a work session, I want to feel like a surgeon walking into a prepped operating room. Everything is ready for me to do what I’m good at.&quot;&lt;/p&gt;</content:encoded><source>Code like a surgeon</source></item><item><title>Do AIs think differently in different languages?</title><link>https://treycausey.com/commonplace/2025-10-17-do-ais-think-differently/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-10-17-do-ais-think-differently/</guid><pubDate>Fri, 17 Oct 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Chatbots are not totally immune to linguistic influence from whichever language they are speaking in, but their worldview does not appear to be strongly determined by language. . . . for the most part, the AIs are secular, Western liberals no matter what language you ask them questions in.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Kelsey Piper with a great piece of empirical AI journalism which basically asks &quot;does the Sapir-Whorf hypothesis hold for LLMs?&quot; It doesn&apos;t really for humans and, it turns out it doesn&apos;t really for LLMs, either! Models are pretty much center-left across the board, with some edge cases specific to DeepSeek when used in Chinese.&lt;/p&gt;
&lt;p&gt;The approach here is a good one, using questions from the World Values Survey, paying translators for questions &amp; answers in Arabic, Hindi, and Chinese, and evaluating multiple state-of-the-art models.&lt;/p&gt;
&lt;p&gt;One of the challenges of making any generalized assertions about how LLMs behave is that most of the academics who are motivated and have the skills to do so aren&apos;t able to move quickly enough to study the newest models. Often, when findings come out, a new generation of models has already emerged.&lt;/p&gt;</content:encoded><source>Do AIs think differently in different languages?</source></item><item><title>Daring Fireball: Matthew Inman of The Oatmeal: &apos;A Cartoonist&apos;s Review of AI Art&apos;</title><link>https://treycausey.com/commonplace/2025-10-16-daringfireball-net-linked-2025-10-16-oatmeal-ai-art/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-10-16-daringfireball-net-linked-2025-10-16-oatmeal-ai-art/</guid><pubDate>Thu, 16 Oct 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;If your opinion of a work art changes after you find out which tools were used to make it, or who the artist is or what they’ve done, you’re no longer judging the art. You’re making a choice not to form your opinion based on the work itself, but rather on something else.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;&quot;Can you separate the art from the artist&quot; is a question that has no definitive answer, with predominant sentiment fluctuating with other political preferences over time. But John Gruber framing it as &quot;a choice&quot; is something I really like.&lt;/p&gt;</content:encoded><source>Daring Fireball: Matthew Inman of The Oatmeal: &apos;A Cartoonist&apos;s Review of AI Art&apos;</source></item><item><title>Technological Optimism and Appropriate Fear</title><link>https://treycausey.com/commonplace/2025-10-13-importai-substack-com-p-import-ai-431-technological-optimism-publication-id-1317673-post-id-176010505-isfreemail-true-r-2ti8-triedredirect-true/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-10-13-importai-substack-com-p-import-ai-431-technological-optimism-publication-id-1317673-post-id-176010505-isfreemail-true-r-2ti8-triedredirect-true/</guid><pubDate>Mon, 13 Oct 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;For us to truly understand what the policy solutions look like, we need to spend a bit less time talking about the specifics of the technology and trying to convince people of our particular views of how it might go wrong - self-improving AI, autonomous systems, cyberweapons, bioweapons, etc. - and more time listening to people and understanding their concerns about the technology. There must be more listening to labor groups, social groups, and religious leaders. The rest of the world which will surely want—and deserves—a vote over this.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Jack Clark shares the remarks he gave at The Curve, the buzzy AI-meets-policy conference (that declined my request to attend 😅).&lt;/p&gt;
&lt;p&gt;While I don&apos;t agree with many of the fears that he shares here, I do have great respect for the earnestness with which he does so.&lt;/p&gt;</content:encoded><source>Technological Optimism and Appropriate Fear</source></item><item><title>In Defense of AI Evals, for Everyone</title><link>https://treycausey.com/commonplace/2025-10-02-www-sh-reya-com-blog-in-defense-ai-evals/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-10-02-www-sh-reya-com-blog-in-defense-ai-evals/</guid><pubDate>Thu, 02 Oct 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;The more interesting question, then, is not whether you do evals, but when you can afford to be less rigorous and when you cannot.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Shreya Shankar patiently -- more patiently than I would do -- responds to the recent evals / no-evals discourse happening (mainly on Twitter). Rather than taking the bait, Shreya assumes good faith and says what this debate is about is really about when it&apos;s ok to be more or less rigorous in your evaluations.&lt;/p&gt;
&lt;p&gt;It&apos;s OK to be less rigorous when your task(s) are already heavily baked into the foundation model&apos;s post-training (such as with coding).&lt;/p&gt;
&lt;p&gt;It&apos;s also ok when you have enough domain expertise and dogfood early and often.&lt;/p&gt;
&lt;p&gt;&lt;blockquote&gt;In my own experience with applications built on top of foundation models (with much less money, lol), evals are especially critical in complex document processing and analysis. Just because a document fits in the context window does not mean the model will complete the task correctly; we have to carefully decompose the task into smaller pieces the model can handle, and then design evals for each of those pieces.&lt;/blockquote&gt;&lt;/p&gt;
&lt;p&gt;In the end, it&apos;s always better to avoid Twitter and form your own conclusions about things instead of parroting the hot takes of the moment.&lt;/p&gt;</content:encoded><source>In Defense of AI Evals, for Everyone</source></item><item><title>Why women should be tech-optimists</title><link>https://treycausey.com/commonplace/2025-10-02-www-theargumentmag-com-p-why-women-should-be-tech-optimists-publication-id-5247799-post-id-175083158-isfreemail-false-r-2ti8-triedredirect-true/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-10-02-www-theargumentmag-com-p-why-women-should-be-tech-optimists-publication-id-5247799-post-id-175083158-isfreemail-false-r-2ti8-triedredirect-true/</guid><pubDate>Thu, 02 Oct 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;A culture of technology that is about freedom rather than domination, about tangible benefits, not abstract achievement — that’s the kind of culture that could turn Waymos from scary and unknown death traps into the liberating bicycles of the 21st century.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Jerusalem Demsas digs into the perplexing but robust finding that women (in 2025, in the US) tend to be more distrustful of and opposed to technological advances despite women also historically being huge beneficiaries of advances.&lt;/p&gt;
&lt;p&gt;This dovetails nicely with &lt;a href=&quot;https://www.slowboring.com/p/capitalism-needs-honor-and-ethics&quot;&gt;today&apos;s Matt Yglesias post&lt;/a&gt; on values and capitalism.&lt;/p&gt;</content:encoded><source>Why women should be tech-optimists</source></item><item><title>10 Things I Hate About AI Evals with Hamel Husain</title><link>https://treycausey.com/commonplace/2025-10-02-www-youtube-com-watch-v-qek-xwrkqhi/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-10-02-www-youtube-com-watch-v-qek-xwrkqhi/</guid><pubDate>Thu, 02 Oct 2025 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;One of the highest signal ~hours I&apos;ve watched recently. Hamel Husain (and his teaching partner Shreya Shankar) are now synonymous with &quot;AI evals&quot;; here he lays out how he thinks people get them wrong and what he&apos;d change.&lt;/p&gt;
&lt;p&gt;Their &lt;a href=&quot;https://maven.com/parlance-labs/evals&quot;&gt;course&lt;/a&gt; is excellent and about to start again.&lt;/p&gt;
&lt;p&gt;Plus, a bonus appearance by Bryan Bischof and the suggestion that &quot;AI evals&quot; should just be renamed &quot;data science for AI&quot;.&lt;/p&gt;
&lt;p&gt;Here are his peeves, highly encourage watching though:&lt;/p&gt;
&lt;p&gt;1) Generic Metrics &amp; Off-the-Shelf Evals
2) Completely Outsourcing Data Review &amp; Leaving Out Domain Experts
3) Overeager Eval Automation
4) Not Looking at the Data At All 
5) Not Thinking Deeply About Prompts
6) Dashboards Full of Noisy Metrics
7) Getting Stuck with Annotation
8) Endlessly Trying Tools Instead of Error Analysis
9) Putting LLMs in the Judge&apos;s Seat Without Human Oversight 
10) Engineers Not Using AI Enough Themselves (Lack of Intuition)&lt;/p&gt;</content:encoded><source>10 Things I Hate About AI Evals with Hamel Husain</source></item><item><title>The algorithm will see you now</title><link>https://treycausey.com/commonplace/2025-10-01-worksinprogress-co-issue-the-algorithm-will-see-you-now/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-10-01-worksinprogress-co-issue-the-algorithm-will-see-you-now/</guid><description>AI isn&apos;t replacing radiologists</description><pubDate>Wed, 01 Oct 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;In many jobs, tasks are diverse, stakes are high, and demand is elastic. When this is the case, we should expect software to initially lead to more human work, not less. The lesson from a decade of radiology models is neither optimism about increased output nor dread about replacement. Models can lift productivity, but their implementation depends on behavior, institutions and incentives.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;An excellent piece from Deena Mousa, lead researcher at OpenPhil, on why radiologists have not been replaced by AI despite the high accuracy and wide availability of radiology-specific models approved for clinical use.&lt;/p&gt;
&lt;p&gt;The real takeaway, for me, though is wondering why people still listen to Hinton et al.&apos;s predictions about the social and economic impact of AI? We need a &quot;bitter lesson&quot; for this kind of thing -- most jobs aren&apos;t just tasks and the implementation of technology into work is complex, political, and sociological.&lt;/p&gt;</content:encoded><source>The algorithm will see you now</source></item><item><title>Responsible AI is dying. Long live responsible AI.</title><link>https://treycausey.com/blog/posts/whither_responsible_ai/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/whither_responsible_ai/</guid><description>How you, yes you, can help build a better future with AI.</description><pubDate>Mon, 29 Sep 2025 00:00:00 GMT</pubDate></item><item><title>Failing to Understand the Exponential, Again</title><link>https://treycausey.com/commonplace/2025-09-29-www-julian-ac-blog-2025-09-27-failing-to-understand-the-exponential-again/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-09-29-www-julian-ac-blog-2025-09-27-failing-to-understand-the-exponential-again/</guid><pubDate>Mon, 29 Sep 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;It may sound overly simplistic, but making predictions by extrapolating straight lines on graphs is likely to give you a better model of the future than most &quot;experts&quot; - even better than most actual domain experts!&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;More on the &quot;AI progress has plateaued&quot; / &quot;AI progress is exploding&quot; debate, this time from Julian Schrittwieser, an AI researcher at Anthropic.&lt;/p&gt;
&lt;p&gt;Reasoning about exponential progress is extremely hard to do for humans. As Schrittwieser points out, we made the same cognitive errors at the beginning of Covid.&lt;/p&gt;
&lt;p&gt;His conclusion: using the METR evals as one measure, we&apos;re right on track and models continue to make exponential progress.&lt;/p&gt;</content:encoded><source>Failing to Understand the Exponential, Again</source></item><item><title>Getting AI to Work in Complex Codebases</title><link>https://treycausey.com/commonplace/2025-09-24-github-com-humanlayer-advanced-context-engineering-for-coding-agents-blob-main-ace-fca-md/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-09-24-github-com-humanlayer-advanced-context-engineering-for-coding-agents-blob-main-ace-fca-md/</guid><pubDate>Wed, 24 Sep 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;The techniques I want to talk about and that we&apos;ve adopted in the last few months fall under what I call &quot;frequent intentional compaction&quot;. Essentially, this means designing your ENTIRE WORKFLOW around context management, and keeping utilization in the 40%-60% range (depends on complexity of the problem ).&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Dex Horthy, founder of &lt;a href=&quot;https://www.hlyr.dev&quot;&gt;HumanLayer&lt;/a&gt;, produces an outstanding guide to getting the results you want from AI when working with large, complex codebases. It&apos;s grounded, reasonable, and eminently actionable with loads of diagrams and examples.&lt;/p&gt;
&lt;p&gt;(Perhaps unsurprisingly) Context is king. Managing your context intentionally and carefully and meticulously is the path to great results. Horthy uses a research - plan - implement pattern that focuses on making sure each step is providing the subsequent step with only the exact context that is needed to be successful.&lt;/p&gt;
&lt;p&gt;It&apos;s really thoughtfully written, I recommend it to everyone, but especially to engineers who have decided that AI isn&apos;t helpful for them.&lt;/p&gt;
&lt;p&gt;See also &lt;a href=&quot;https://x.com/ericzakariasson/status/1969194919597392271&quot;&gt;this short context engineering template from Eric Zakariasson&lt;/a&gt; that arrives at similar conclusions.&lt;/p&gt;</content:encoded><source>Getting AI to Work in Complex Codebases</source></item><item><title>A Billion Dollars isn&apos;t cool. You know what is? Ten million dollars.</title><link>https://treycausey.com/blog/posts/ai_business_models/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/ai_business_models/</guid><description>AI can revive the middle class business model</description><pubDate>Tue, 23 Sep 2025 00:00:00 GMT</pubDate></item><item><title>Thinking, Searching, and Acting</title><link>https://treycausey.com/commonplace/2025-09-22-www-interconnects-ai-p-thinking-searching-and-acting-publication-id-48206-post-id-174200501-isfreemail-false-r-2ti8-triedredirect-true/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-09-22-www-interconnects-ai-p-thinking-searching-and-acting-publication-id-48206-post-id-174200501-isfreemail-false-r-2ti8-triedredirect-true/</guid><pubDate>Mon, 22 Sep 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;With search, hallucinations are now missing context rather than blatantly incorrect content. Language models are nearly-perfect at copying content and similarly solid at referencing it, but they&apos;re still very flawed at long-context understanding. Hallucinations still matter, but it’s a very different chapter of the story and will be studied differently depending on if it is for reasoning or non-reasoning language models.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Nathan Lambert of Ai2 provides a good summary of what constitutes a &apos;reasoning model&apos; such as OpenAI&apos;s o3 or GPT-5 Thinking. Like much of the fast-moving AI field, common definitions are hard to come by, and clear definitions are even rarer.&lt;/p&gt;
&lt;p&gt;Before these models were released, models were effectively limited to &quot;just&quot; next-token prediction based on the information in their pretraining data and with a hard knowledge cut-off date.&lt;/p&gt;
&lt;p&gt;While base models still predict next tokens and knowledge cut-off dates are still very much a real thing, reasoning models can search and obtain new information needed to complete tasks.&lt;/p&gt;
&lt;p&gt;Lambert breaks down reasoning models into three primitives that distinguish them from their static predecessors: thinking (i.e., dynamically using more tokens to arrive at an answer), searching, and acting / executing.&lt;/p&gt;</content:encoded><source>Thinking, Searching, and Acting</source></item><item><title>You Had No Taste Before AI</title><link>https://treycausey.com/commonplace/2025-09-21-matthewsanabria-dev-posts-you-had-no-taste-before-ai/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-09-21-matthewsanabria-dev-posts-you-had-no-taste-before-ai/</guid><pubDate>Sun, 21 Sep 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;To play on a popular quote from Ratatouille, anyone can cook, but not everyone is a chef. Don’t complain about mediocre work when you’re producing mediocre work yourself.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Matthew Sanabria encourages you, yes you, to actually develop taste or refine your taste instead of lamenting the tastelessness of AI outputs. As the title claims, many people had no taste before, they&apos;re just now realizing that there&apos;s lots of bad, tasteless content out there.&lt;/p&gt;
&lt;p&gt;I&apos;ve been noticing this as well. I have read / heard countless variants of &quot;I&apos;ve never read anything well-written by AI.&quot; Besides the fact that this is really a claim about &lt;em&gt;obviously&lt;/em&gt; AI-written text (which can be quite banal), it makes me wonder how many people consider themselves truly excellent writers.&lt;/p&gt;
&lt;p&gt;Before the advent of LLMs, it seemed to be a truism that writing well was an extremely difficult, rare skill. Good writers received accolades from their peers. In my mind, very few people considered themselves excellent writers.&lt;/p&gt;
&lt;p&gt;Is everyone an excellent writer all of a sudden? That seems implausible.&lt;/p&gt;</content:encoded><source>You Had No Taste Before AI</source></item><item><title>Is AI a bubble?</title><link>https://treycausey.com/commonplace/2025-09-17-www-exponentialview-co-p-is-ai-a-bubble-publication-id-2252-post-id-173831079-isfreemail-false-r-2ti8-triedredirect-true/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-09-17-www-exponentialview-co-p-is-ai-a-bubble-publication-id-2252-post-id-173831079-isfreemail-false-r-2ti8-triedredirect-true/</guid><pubDate>Wed, 17 Sep 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;On the basis of these gauges, genAI remains in a demand-led, capital-intensive boom rather than a bubble. But booms can sour quickly, and there are several pressure points worth watching&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Azeem Azhar and Nathan Warren do some in-the-weeds research to try and answer what is a seemingly straightforward question -- is AI a bubble?&lt;/p&gt;
&lt;p&gt;The truth is that bubbles are only recognizable in retrospect; what may appear to be a &quot;boom&quot; has to &quot;burst&quot; in order to be a bubble. However, identifying signs that might indicate that a boom is likely to be a bursting bubble is extremely valuable.&lt;/p&gt;
&lt;p&gt;Acknowledging this, the authors cover a lot of ground and come up with five factors to watch: economic strain (capex / GDP %), industry strain (investment / revenue), revenue growth (revenue doubling time in years), valuation heat (price / earnings ratio), and funding quality (a composite index).&lt;/p&gt;
&lt;p&gt;Their conclusion -- AI is still firmly in &quot;boomb&quot; territory, rather than looking like a bubble. Helpfully, they also provide a few warning signs to watch for: if investment in AI approaches 2% of GDP, sustained drops in enterprise and consumer spending, P/E ratios approaching 50-60, and if internal cash covers less than 25% of capex. Two of these would push the authors into the danger zone.&lt;/p&gt;
&lt;p&gt;Is this a perfect study? Probably not. Is it possible they&apos;ve misread the data? Certainly. But what is certainly true is they&apos;ve taken the question seriously, looked for unbiased indicators, tested their validity qualitatively and quantitatively, and even shared how they might be wrong.&lt;/p&gt;
&lt;p&gt;We need a &lt;em&gt;lot&lt;/em&gt; more of this and a lot less of the smug, overconfident, tweet-length assertions that grab most of the headlines today.&lt;/p&gt;</content:encoded><source>Is AI a bubble?</source></item><item><title>Why Language Models Hallucinate</title><link>https://treycausey.com/commonplace/2025-09-11-cdn-openai-com-pdf-d04913be-3f6f-4d2b-b283-ff432ef4aaa5-why-language-models-hallucinate-pdf/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-09-11-cdn-openai-com-pdf-d04913be-3f6f-4d2b-b283-ff432ef4aaa5-why-language-models-hallucinate-pdf/</guid><pubDate>Thu, 11 Sep 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;&quot;[L]anguage models hallucinate because the training and evaluation procedures reward guessing over acknowledging uncertainty&quot;&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;&quot;Why Language Models Hallucinate&quot;, released by Kalai, Nachum, Vempala, and Zhang (mostly of OpenAI) last week.&lt;/p&gt;
&lt;p&gt;It has both a potentially surprising conclusion and a definitely (to me) surprising recommendation for solutions -- surprising to me because I made a similar point a couple of months ago!&lt;/p&gt;
&lt;p&gt;The tl;dr for the conclusion is that language models hallucinate because of the way they are trained and evaluated. No, not because of next-token prediction necessarily, but because of the incentives to provide an answer. Language models are &quot;optimized to be good test-takers&quot; where saying &quot;I don&apos;t know&quot; is not rewarded.&lt;/p&gt;
&lt;p&gt;The authors demonstrate that, largely independent of architecture, even in situations where only correct factual information is present in the training data models will still be incentivized to guess incorrectly.&lt;/p&gt;
&lt;p&gt;The proposed solution is (naturally) socio-technical by changing how models are scored and disincentivizing guessing over expressing uncertainty.&lt;/p&gt;
&lt;p&gt;See: &lt;a href=&quot;https://www.treycausey.com/blog/posts/solving_llm_hallucinations_ux/&quot;&gt;Solving LLM Hallucinations is (mostly) a UX Problem&quot;&lt;/a&gt;&lt;/p&gt;</content:encoded><source>Why Language Models Hallucinate</source></item><item><title>Quoting Koen</title><link>https://treycausey.com/commonplace/2025-09-07-x-com-koenvaneijk-status-1964545137784340736/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-09-07-x-com-koenvaneijk-status-1964545137784340736/</guid><pubDate>Sun, 07 Sep 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;treat your code like bonsai, ai makes it grow faster, prune it from time to time to keep structure and establish its form.&lt;/p&gt;&lt;/blockquote&gt;</content:encoded><source>Quoting Koen</source></item><item><title>Quoting Anil Dash on AI critiques</title><link>https://treycausey.com/commonplace/2025-09-07-bsky-app-profile-anildash-com-post-3lybkmj7ast2c/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-09-07-bsky-app-profile-anildash-com-post-3lybkmj7ast2c/</guid><pubDate>Sun, 07 Sep 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;[T]his is the really ineffective head-in-the-sand reaction a lot of folks on Bluesky tend to have if you talk about any harm reduction-based approach to the reality that millions of people are using AI tools today. The current approach to critiquing AI is obviously not working, but they don&apos;t care. ... I agree with the intellectual substance of virtually every common critique of AI. And it&apos;s very clear that turning those critiques into a competition about who can frame them in the most scathing way online has done *zero* to slow down adoption, even if much of that is due to default bundling. At what point are folks going to try literally any other tactic than condescending rants? Does it matter that LLM apps are at the top of virtually every app store nearly every day because individual people are choosing to download them, and the criticism hasn&apos;t been effective in slowing that?&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Anil Dash on Bluesky. I couldn&apos;t agree with this more. I have a longer post brewing on where I think the field of &quot;Responsible AI&quot; (vs. the idea of developing AI responsibly) is heading, but it&apos;s not quite ready for prime time yet.&lt;/p&gt;</content:encoded><source>Quoting Anil Dash on AI critiques</source></item><item><title>AI IS SLOWING DOWN Tracker</title><link>https://treycausey.com/commonplace/2025-08-29-aislowdown-replit-app/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-29-aislowdown-replit-app/</guid><pubDate>Fri, 29 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;&quot;AI progress is slowing down&quot; is a meme with surprising legs, especially after the polarizing GPT-5 launch. Peter Gostev uses that same GPT-5 to create this tracker of media stories and articles dating to early 2023 decrying the end of AI progress. I love this and am more than a little irritated I didn&apos;t think of it first.&lt;/p&gt;</content:encoded><source>AI IS SLOWING DOWN Tracker</source></item><item><title>Measuring the environmental impact of AI inference</title><link>https://treycausey.com/commonplace/2025-08-28-cloud-google-com-blog-products-infrastructure-measuring-the-environmental-impact-of-ai-inference/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-28-cloud-google-com-blog-products-infrastructure-measuring-the-environmental-impact-of-ai-inference/</guid><pubDate>Thu, 28 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;[W]e estimate the median Gemini Apps text prompt uses 0.24 watt-hours (Wh) of energy, emits 0.03 grams of carbon dioxide equivalent (gCO2e), and consumes 0.26 milliliters (or about five drops) of water1 — figures that are substantially lower than many public estimates. The per-prompt energy impact is equivalent to watching TV for less than nine seconds.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Google releases environmental impact data for Gemini. Much like the Mistral report previously released, the impact is significantly smaller than popular discourse would lead you to believe. This is really encouraging, but unless the discourse catches up, it&apos;s hard to see it moving the needle.&lt;/p&gt;</content:encoded><source>Measuring the environmental impact of AI inference</source></item><item><title>&quot;For All Issues So Triable&quot;</title><link>https://treycausey.com/commonplace/2025-08-28-www-hyperdimensional-co-p-for-all-issues-so-triable-utm-source-post-email-title-publication-id-2244049-post-id-172122376-utm-campaign-email-post-title-isfreemail-true-r-2ti8-triedredirect-true-utm-medium-email/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-28-www-hyperdimensional-co-p-for-all-issues-so-triable-utm-source-post-email-title-publication-id-2244049-post-id-172122376-utm-campaign-email-post-title-isfreemail-true-r-2ti8-triedredirect-true-utm-medium-email/</guid><pubDate>Thu, 28 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;And for all the millions of dollars spent on AI safety non-profits, advocacy organizations, and other research efforts, almost none of it has been devoted specifically to the issues implicated in Raine. Indeed, we have more robust methods for measuring and mitigating LLM-enabled bioweapon development risk than we do for the rather more mundane, but far more visceral, issue of how a chatbot should deal with a mentally troubled teenager. This critique applies to my own writing as well.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Dean Ball, one of the authors of the US AI Action Plan just released by the Trump administration, weighs in on &lt;em&gt;Raine v. OpenAI&lt;/em&gt; and its potential to be a landmark case that defines how individuals seek redress against AI labs.&lt;/p&gt;</content:encoded><source>&quot;For All Issues So Triable&quot;</source></item><item><title>Quoting Derek Thompson</title><link>https://treycausey.com/commonplace/2025-08-27-www-derekthompson-org-p-the-evidence-that-ai-is-destroying/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-27-www-derekthompson-org-p-the-evidence-that-ai-is-destroying/</guid><pubDate>Wed, 27 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Someone once asked me recently if I had any advice on how to predict the future when I wrote about social and technological trends. Sure, I said. My advice is that predicting the future is impossible, so the best thing you can do is try to describe the present accurately. Since most people live in the past, hanging onto stale narratives and outdated models, people who pay attention to what’s happening as it happens will appear to others like they’re predicting the future when all they’re doing is describing the present.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;In his (subscriber-only) discussion of emerging evidence that AI &lt;del&gt;is&lt;/del&gt; &lt;del&gt;isn&apos;t&lt;/del&gt; might be negatively affecting early career job prospects, Derek Thompson drops this great kicker.&lt;/p&gt;</content:encoded><source>Quoting Derek Thompson</source></item><item><title>The AI Productivity Curve.</title><link>https://treycausey.com/commonplace/2025-08-25-writing-pupius-co-uk-the-ai-productivity-curve-8a426646cf64/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-25-writing-pupius-co-uk-the-ai-productivity-curve-8a426646cf64/</guid><pubDate>Mon, 25 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Productivity has always been a probability function — sometimes you’re in flow, sometimes you’re stuck for hours. AI changes the shape of that distribution. Understanding how is more useful than arguing about how much.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Dan Pupius with a nuanced view on why the data / anecdata on coding productivity gains from AI are all over the place. For him, the biggest gains are in getting started and flattening as complexity grows. This seems to me a) both obviously true and b) the first place I&apos;m seeing it articulated so clearly. The discussion around AI and productivity has sadly become, like almost every other AI discussion, effectively a litmus test for your priors about the pros and cons of AI.&lt;/p&gt;
&lt;p&gt;Related: &lt;a href=&quot;https://www.treycausey.com/blog/posts/superagency-adhd/&quot;&gt;Superagency and ADHD&lt;/a&gt;&lt;/p&gt;</content:encoded><source>The AI Productivity Curve</source></item><item><title>On Pessimization</title><link>https://treycausey.com/commonplace/2025-08-20-www-mindthefuture-info-p-on-pessimization/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-20-www-mindthefuture-info-p-on-pessimization/</guid><pubDate>Wed, 20 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;&quot;The more scared you are that you might not achieve your goal, the more urgently you feel that “something must be done”, and the more you flinch away from picturing how that “something” might actually make it worse.&quot;&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Richard Ngo, an AI safety advocate formerly with Google DeepMind and OpenAI, writes a really lovely piece about &quot;pessimization&quot;, the phenomenon where working towards some goal counterintuitively helps to achieve the opposite of that goal (i.e., an anti-goal).&lt;/p&gt;
&lt;p&gt;I love this kind of concept; they appear all over the social sciences with various names. They are a vital part of examining how systems and institutions and the actors within them operate, but are woefully underdiscussed outside of the academy. Even when they are discussed, in my experience, they are treated as interesting thought experiments or narrow applications. In reality, I think they&apos;re far more common that acknowledged.&lt;/p&gt;
&lt;p&gt;Ngo breaks down pessimization into a three-part typology:
&lt;ul&gt;
	&lt;li /&gt;Direct pessimization: when opponents actively try to bring about the anti-goal.
	&lt;li /&gt;Indirect pessimization: when work towards a goal encourages others to work on the anti-goal either through raising awareness of the anti-goal or simply by having bad or uncompelling ideas about the goal itself.
	&lt;li /&gt;Perverse pessimization: when ostensibly aligned actors sabotage progress towards the goal. This is common when organizational and institutional structures emerge in support of a goal whose existence would be threatened if the goal was realized. Ngo cites the prime example here of environmentalists opposing nuclear power.
&lt;/ul&gt;&lt;/p&gt;
&lt;p&gt;One of the feedback effects that Ngo doesn&apos;t discuss, and one I think a lot about, is something akin to but not quite equivalent to the No True Scotsman fallacy. The idea being here that, in the face of failure to achieve a goal, the proposed solution is to double down on existing methods and tactics rather than admit that they are not succeeding. &quot;Real \&lt;approach\&gt; has never been tried, that&apos;s why we&apos;re failing!&quot;&lt;/p&gt;</content:encoded><source>On Pessimization - by Richard Ngo - Mind the Future</source></item><item><title>OpenAI Progress over 7 years</title><link>https://treycausey.com/commonplace/2025-08-17-progress-openai-com/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-17-progress-openai-com/</guid><pubDate>Sun, 17 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;I&apos;ve wanted to do this for a while. OpenAI feeds the same prompt (well, 14 of them) to five different LLMs ranging from 2018&apos;s GPT-1 to 2025&apos;s GPT-5. Pretty remarkable progress.&lt;/p&gt;</content:encoded><source>OpenAI Progress</source></item><item><title>Different perceptions of AI progress, explained</title><link>https://treycausey.com/commonplace/2025-08-17-x-com-miles-brundage-status-1956777848523641016/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-17-x-com-miles-brundage-status-1956777848523641016/</guid><pubDate>Sun, 17 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;Miles Brundage, former Head of Policy at OpenAI, walks through why there are so many competing narratives about AI progress right now. His conclusion is that many people are talking about different things altogether when they talk about &quot;progress&quot; and why it can feel simultaneously like both narratives are true: that progress has plateaued and that new models continue to break new ground. It&apos;s a good six-minute overview of the different kinds of models that are out there without getting too technical about architectures.&lt;/p&gt;</content:encoded><source>Different perceptions of AI progress, explained</source></item><item><title>AI research interviews</title><link>https://treycausey.com/commonplace/2025-08-13-docs-google-com-document-d-1zv73d2vgaj2yu-tjn3tvop6qvlwvpxjb2rrqszqxyti-edit/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-13-docs-google-com-document-d-1zv73d2vgaj2yu-tjn3tvop6qvlwvpxjb2rrqszqxyti-edit/</guid><pubDate>Wed, 13 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;As an order of magnitude, I’d recommend around 100 hours of leetcode practice, and a similar amount of time reading papers, refreshing knowledge (use Deep Research!) and talking to friends.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Bas van Opheusden, recent addition to the technical staff at OpenAI, shares his advice for preparing for AI research lab interviews. It pains me that we occupy an equilibrium where leetcoding is broadly seen as suboptimal for identifying talent yet is the dominant form of technical interview. &lt;a href =&quot;https://en.wikipedia.org/wiki/Streetlight_effect&quot;&gt;Classic streetlight effect&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;These guides can often feel like advice to &quot;know everything about everything&quot; (I should know, I wrote one myself some years ago) but there are some good nuggets:&lt;/p&gt;
&lt;p&gt;Concrete interview topics to prepare:&lt;/p&gt;
&lt;p&gt;&lt;ul&gt;
	&lt;li /&gt;Debugging transformers. This is a classic in which you get a botched implementation of a self-attention block and have to debug it. Make sure you’ve practiced debugging tensor shapes, and pay special attention to the causal attention mask - that’s where it gets most tricky. For reference, check out nanoGPT. 
	&lt;li /&gt;Top-k/knn. The problem of “picking the k largest items” comes up in ML in various places and makes for a nice interview problem, particularly because the solution (a heap) is not something you can invent on-the-fly. Make sure you know what heaps are. 
	&lt;li /&gt;Implementing BPE. Tokenizers are the worst part of LLMs, and implementing BPE without errors is tricky. This is somewhat popular. 
	&lt;li /&gt;Backpropagation from scratch. Implementing a basic version of auto-diff, the chain rule etc. Lots of opportunities for indexing errors.  
	&lt;li /&gt;KV Cache. This essentially amounts to building a matrix, but if you haven’t seen it before, you might do something convoluted 
	&lt;li /&gt;Binary search, backtracking, dijkstra, ...
&lt;/ul&gt;&lt;/p&gt;</content:encoded><source>AI research interviews</source></item><item><title>42 notes on AI &amp; work by Jasmine Sun</title><link>https://treycausey.com/commonplace/2025-08-13-substack-com-jasmine-p-170754042/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-13-substack-com-jasmine-p-170754042/</guid><pubDate>Wed, 13 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;I don’t think policymakers would tolerate job loss past 15%. At that point, they’d step in to start slowing shit down.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;I love lists like these. Jasmine Sun collects a series of thoughts / observations / datapoints from the past few years about the intersection of AI and work. Sometimes contradictory -- that&apos;s the point -- and definitely evidence of continuing to learn and grapple with a complex topic.&lt;/p&gt;
&lt;p&gt;Many of these had me nodding my head, and Jasmine seems to be one of the only other people besides Noah Smith living in SF, immersed in the AI culture, hanging out with the builders, who takes seriously the real political, economic, and institutional realities in which AI is being developed.&lt;/p&gt;</content:encoded><source>42 notes on AI &amp; work by Jasmine Sun</source></item><item><title>The ‘godfather of AI’ reveals the only way humanity can survive superintelligent AI</title><link>https://treycausey.com/commonplace/2025-08-13-www-cnn-com-2025-08-13-tech-ai-geoffrey-hinton/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-13-www-cnn-com-2025-08-13-tech-ai-geoffrey-hinton/</guid><pubDate>Wed, 13 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;“That’s the only good outcome. If it’s not going to parent me, it’s going to replace me,” he said. “These super-intelligent caring AI mothers, most of them won’t want to get rid of the maternal instinct because they don’t want us to die.”&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Geoffrey Hinton has now decided that building &quot;maternal instincts&quot; into AI is the only way to prevent all of humanity being wiped out. My hope is that more people read this kind of thing and come to the correct conclusion that computer scientists are not well-equipped to make sweeping predictions about social systems.&lt;/p&gt;
&lt;p&gt;The quotes are ... something else.&lt;/p&gt;</content:encoded><source>The ‘godfather of AI’ reveals the only way humanity can survive superintelligent AI</source></item><item><title>Thrive in obscurity</title><link>https://treycausey.com/commonplace/2025-08-11-www-jeetmehta-com-posts-thrive-in-obscurity/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-11-www-jeetmehta-com-posts-thrive-in-obscurity/</guid><pubDate>Mon, 11 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Do things you like, and sometimes the world will agree&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Jeet Mehta shares three principles for creating things when you don&apos;t have an audience:&lt;/p&gt;
&lt;p&gt;&lt;ol&gt;
	&lt;li /&gt;Do things you like, and sometimes the world will agree.
	&lt;li /&gt;Push yourself out: create for yourself, not for an imagined audience.
	&lt;li /&gt;Build your binge bank: consider your unread writing a bingeable backlog for those who discover you.
&lt;/ol&gt;&lt;/p&gt;
&lt;p&gt;I think 1 &amp; 3 contradict a bit, but overall I like the message a lot. Jeet also posted his &lt;a href=&quot;https://www.jeetmehta.com/posts/core-beliefs&quot;&gt;core beliefs&lt;/a&gt;, which I really admire.&lt;/p&gt;</content:encoded><source>Thrive in obscurity</source></item><item><title>Amazon deliveries are often better for the environment than driving to buy the stuff in a store</title><link>https://treycausey.com/commonplace/2025-08-08-andymasley-substack-com-p-ordering-amazon-deliveries-is-often/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-08-andymasley-substack-com-p-ordering-amazon-deliveries-is-often/</guid><pubDate>Fri, 08 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;If you often drive to buy things, getting things delivered to your house will often be better for the climate, unless you’re buying a ton of stuff at once at a store that’s close by. Driving is just so uniquely bad for the climate compared to everything else we do that we shouldn’t be surprised by this.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Andy Masley outlines the environmental impact of getting things delivered to your door from Amazon vs. driving to the store to do so.&lt;/p&gt;
&lt;p&gt;A few things are at work here, I think:&lt;/p&gt;
&lt;p&gt;&lt;ul&gt;
	&lt;li /&gt;Most people never really seriously consider conterfactuals.
	&lt;li /&gt;The efficiency of modern logistics and supply chains is vastly underestimated (see also discussions about eating local).
	&lt;li /&gt;Climate claims that take seriously logistical efficiency are not aesthetically aligned with the preferences of many inclined to make climate claims.
	&lt;li /&gt;(🔥) Climate arguments are a socially acceptable way to criticize things that are also disliked for non-climate reasons.
&lt;/ul&gt;&lt;/p&gt;
&lt;p&gt;On the last point, also see Masley&apos;s piece &lt;a href=&quot;https://andymasley.substack.com/p/a-cheat-sheet-for-conversations-about&quot;&gt;Why using ChatGPT is not bad for the environment - a cheat sheet&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;Related: &lt;a href=&quot;https://www.treycausey.com/blog/posts/quantification_bias/&quot;&gt;Quantification bias&lt;/a&gt;.&lt;/p&gt;</content:encoded><source>Amazon deliveries are often better for the environment than driving to buy the stuff in a store</source></item><item><title>Jerry Wei on one year at Anthropic</title><link>https://treycausey.com/commonplace/2025-08-08-x-com-jerryweiai-status-1932865074361975151/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-08-x-com-jerryweiai-status-1932865074361975151/</guid><pubDate>Fri, 08 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Being surrounded by colleagues who embody these principles, who choose impact over recognition, who obsess over measurement quality, and who believe in the power of focused teams, has shaped how I approach my own work.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Jerry Wei reflects on one year at Anthropic. As with most things that are true, they sound &quot;obvious&quot; but are often quite difficult to keep in mind when trying to build successful teams.&lt;/p&gt;
&lt;p&gt;&lt;ul&gt;
	&lt;li /&gt;Small, talent-dense, goal-aligned teams can outpace much larger teams.
	&lt;li /&gt;Continually refining evals and making them harder and more comprehensive is essential to success.
	&lt;li /&gt;The highest impact work often isn&apos;t the most glamorous.
&lt;/ul&gt;&lt;/p&gt;</content:encoded><source>Twitter</source></item><item><title>A Spark of the Anti-AI Butlerian Jihad (on Bluesky)</title><link>https://treycausey.com/commonplace/2025-08-03-eugeneyan-com-writing-anti/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-03-eugeneyan-com-writing-anti/</guid><pubDate>Sun, 03 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Despite the unpleasant events, I understand it comes from a place of concern and fear for many. To alleviate this, we need open, empathetic dialog to build mutual understanding. Even now, several AI researchers and practitioners continue to engage in good-faith education about AI on Bluesky. Nonetheless, it’s unclear if critics are open to engage, especially with the widespread blocking of AI-associated accounts.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Eugene Yan finds that it&apos;s not just vibes, Bluesky appears to be a particularly anti-AI platform. What&apos;s nice about this post is it lays out how unwillingness to engage reinforces misconceptions about AI and open data. Polarization is everywhere.&lt;/p&gt;</content:encoded><source>A Spark of the Anti-AI Butlerian Jihad (on Bluesky)</source></item><item><title>Stevens: a hackable AI assistant using a single SQLite table and a handful of cron jobs</title><link>https://treycausey.com/commonplace/2025-08-02-www-geoffreylitt-com-2025-04-12-how-i-made-a-useful-ai-assistant-with-one-sqlite-table-and-a-handful-of-cron-jobs/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-02-www-geoffreylitt-com-2025-04-12-how-i-made-a-useful-ai-assistant-with-one-sqlite-table-and-a-handful-of-cron-jobs/</guid><pubDate>Sat, 02 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Vibe coding enables sillier projects. Initially, Stevens spoke with a dry tone, like you might expect from a generic Apple or Google product. But it turned out it was just more fun to have the assistant speak like a formal butler. This was trivial to do, just a couple lines in a prompt. Similarly, I decided to make the admin dashboard views feel like a video game, because why not?&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Geoffrey Litt, of Ink &amp; Switch, creates a dead-simple, Telegram-based LLM assistant for him and his wife. Impressive how far he gets with a SQLite table for memory and some cron jobs.&lt;/p&gt;</content:encoded><source>Stevens: a hackable AI assistant using a single SQLite table and a handful of cron jobs</source></item><item><title>A thread of non-code Claude Code use cases</title><link>https://treycausey.com/commonplace/2025-08-02-x-com-alexalbert-status-1948765443776544885/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-08-02-x-com-alexalbert-status-1948765443776544885/</guid><pubDate>Sat, 02 Aug 2025 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;Includes a &lt;a href=&quot;https://t.co/mNYiubbkp2&quot;&gt;particularly interesting use case&lt;/a&gt; using CC to organize Obsidian vaults.&lt;/p&gt;</content:encoded><source>A thread of non-code Claude Code use cases</source></item><item><title>Superagency and ADHD</title><link>https://treycausey.com/blog/posts/superagency-adhd/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/superagency-adhd/</guid><description>Imagining a world look like where more people can just do things</description><pubDate>Fri, 01 Aug 2025 00:00:00 GMT</pubDate></item><item><title>Anthropic is launching an AI Psychiatry team</title><link>https://treycausey.com/commonplace/2025-07-26-www-reddit-com-r-artificial-comments-1m8uon6-anthropic-is-launching-an-ai-psychiatry-team-to/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-07-26-www-reddit-com-r-artificial-comments-1m8uon6-anthropic-is-launching-an-ai-psychiatry-team-to/</guid><description>Models contain unknown multitudes</description><pubDate>Sat, 26 Jul 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;We&apos;re launching an &quot;AI psychiatry&quot; team as part of interpretability efforts at Anthropic!  We&apos;ll be researching phenomena like model personas, motivations, and situational awareness, and how they lead to spooky/unhinged behaviors.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Jack Lindsey of Anthropic announces an &quot;AI psychiatry&quot; team at Anthropic. A bit cheeky, as the Jack Lindsey of Anthropic announces an &quot;AI psychiatry&quot; team at Anthropic. A bit cheeky, as the 
&lt;a href=&apos;https://job-boards.greenhouse.io/anthropic/jobs/4020159008&apos;&gt;linked job&lt;/a&gt; doesn&apos;t mention this phrase at all and instead identifies the role as part of Anthropic&apos;s Interpretability team.&lt;/p&gt;
&lt;p&gt;That being said, there is clearly a large and unexplored world of weird model behaviors: [an anon on Twitter discusses](https://x.com/AITechnoPagan/status/1948578422361297321) &apos;Cat Mode&apos; that they discovered within Bing, and the [Claude 4 System Card](https://www-cdn.anthropic.com/6be99a52cb68eb70eb9572b4cafad13df32ed995.pdf) itself discusses an odd &quot;&apos;spiritual bliss&apos; attractor state&quot; (5.5.2, page 62).&lt;/p&gt;</content:encoded><source>Anthropic is launching an AI Psychiatry team</source></item><item><title>Our contribution to a global environmental standard for AI | Mistral AI</title><link>https://treycausey.com/commonplace/2025-07-24-mistral-ai-news-our-contribution-to-a-global-environmental-standard-for-ai/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-07-24-mistral-ai-news-our-contribution-to-a-global-environmental-standard-for-ai/</guid><pubDate>Thu, 24 Jul 2025 00:00:00 GMT</pubDate><content:encoded>&lt;p&gt;Mistral releases the most detailed, independently audited environmental impact assessment of any LLM I&apos;ve seen, their Large 2 (roughly equivalent to GPT-4o in capabilities). With any assessment, whether the impact is &quot;big&quot; or &quot;bad&quot; really depends on the frame you&apos;re using. They break out the impact between pretraining and inference, as is sensible, and give some comparison activities. For the average user, I think the news is quite heartening, a single query uses 0.05L of water (roughly 3 tablespoons) and emits the same amount of greenhouse gases as about 10 seconds of watching a livestream.&lt;/p&gt;
&lt;p&gt;Obviously, inference is much costlier, but also is something that the typical user has almost no control over and is also a one-time fixed cost vs. the ongoing cost of inference. I expect these numbers to continue to fall.&lt;/p&gt;
&lt;p&gt;![Mistral AI&apos;s Large 2 Environmental Impact](../../images/mistral_environmental_impact.png)&lt;/p&gt;</content:encoded><source>Our contribution to a global environmental standard for AI | Mistral AI</source></item><item><title>AI as the greatest source of empowerment for all</title><link>https://treycausey.com/commonplace/2025-07-22-openai-com-index-ai-as-the-greatest-source-of-empowerment-for-all/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-07-22-openai-com-index-ai-as-the-greatest-source-of-empowerment-for-all/</guid><description>Fidji Simo&apos;s guardedly optimistic view of AI</description><pubDate>Tue, 22 Jul 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Every major technology shift can expand access to power—the power to make better decisions, shape the world around us, and control our own destiny in new ways. But it can also further concentrate wealth and power in the hands of a few—usually people who already have money, credentials, and connections.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Fidji Simo, incoming CEO of Applications for OpenAI, lays out her vision for the intentional development of AI across knowledge, health, creative expression, economic freedom, time, and support. A solid opening statement and a refreshing acknowledgement of the work we must do to move towards an abundant future for all.&lt;/p&gt;</content:encoded><source>AI as the greatest source of empowerment for all</source></item><item><title>Stop pretending you know what AI does to the economy</title><link>https://treycausey.com/commonplace/2025-07-21-www-noahpinion-blog-p-stop-pretending-you-know-what-ai/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-07-21-www-noahpinion-blog-p-stop-pretending-you-know-what-ai/</guid><pubDate>Mon, 21 Jul 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;AI might start killing jobs en masse and sending inequality to the moon. We don’t know. But it hasn’t yet, and it’s important to understand why each burst of AI pessimism so far has been a false alarm.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Noah Smith with the most level-headed assessment I&apos;ve read so far about the seemingly omnipresent assumption that AI will be a massive job-killer and the fact that this isn&apos;t visible in any data (so far).&lt;/p&gt;
&lt;p&gt;I think, in general, people are afraid to say &quot;I don&apos;t know.&quot; I have a hypothesis that &quot;I don&apos;t know&quot; is essentially coded as agreeing with whatever outcome is less politically or culturally aligned. I have no data for this, though.&lt;/p&gt;</content:encoded><source>Stop pretending you know what AI does to the economy</source></item><item><title>Notes on running a link blog</title><link>https://treycausey.com/commonplace/2025-07-20-notes-on-running-a-link-blog-/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-07-20-notes-on-running-a-link-blog-/</guid><pubDate>Sun, 20 Jul 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;The point of that article was to emphasize that blogging doesn’t have to be about unique insights. The value is in writing frequently and having something to show for it over time—worthwhile even if you don’t attract much of an audience (or any audience at all).&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;Simon Willison&apos;s guidelines for running his link blog. I&apos;m trying to start doing more of the same, and these seem well-considered.&lt;/p&gt;
&lt;p&gt;- Include the names of people involved (for credit and for discovery of connections to other projects)
- Try to add something extra -- protect against link rot!
- Context as to why interesting / important
- Link to similar concepts
- Quote the transcript if a video
- Original author should feel good about it
- Prove you&apos;ve read it&lt;/p&gt;</content:encoded><source>My approach to running a link blog</source></item><item><title>Notes on Notes on &apos;Taste&apos;</title><link>https://treycausey.com/commonplace/2025-07-20-notes-on-taste-are-na-editorial/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-07-20-notes-on-taste-are-na-editorial/</guid><pubDate>Sun, 20 Jul 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;I also believe taste is something we can and should try to cultivate. Not because taste itself is a virtue, per se, but because I’ve found a taste-filled life to be a richer one. To pursue it is to appreciate ourselves, each other, and the stuff we’re surrounded by a whole lot more.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;The universe kept nudging me towards this, after reading Brie Wolfson&apos;s profoundly affecting &lt;a href=&quot;https://joincolossus.com/article/flounder-mode/&quot;&gt;Flounder Mode&lt;/a&gt;, and her mentioning it on both Packy McCormick&apos;s podcast and David Perell&apos;s.&lt;/p&gt;
&lt;p&gt;The word &apos;taste&apos; is having a moment, with those in the AI research community commonly referring to &apos;research taste&apos; (not sure if Chris Olah invented the term, but &lt;a href=&quot;https://colah.github.io/notes/taste/&quot;&gt;he discussed it in 2021&lt;/a&gt;).&lt;/p&gt;
&lt;p&gt;Elsewhere in the AI community, the importance of taste is a hot topic as the cost of producing endless variations on an idea goes to ~zero.&lt;/p&gt;</content:encoded><source>Notes on &apos;Taste&apos;</source></item><item><title>Solving LLM hallucinations is (mostly) a UX problem</title><link>https://treycausey.com/blog/posts/solving_llm_hallucinations_ux/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/solving_llm_hallucinations_ux/</guid><description>Why LLM hallucinations are primarily a UX challenge rather than an algorithmic one, and how design improvements can mitigate their impact on users.</description><pubDate>Tue, 08 Jul 2025 00:00:00 GMT</pubDate></item><item><title>Letter to Arc members 2025</title><link>https://treycausey.com/commonplace/2025-05-28-letter-to-arc-members-2025-/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-05-28-letter-to-arc-members-2025-/</guid><description>Arc abandoned for Dia—AI moment too big to ignore</description><pubDate>Wed, 28 May 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Second, I would’ve embraced AI fully, sooner and unapologetically. The truth is I was obsessed. I’d stay up late, after my family went to bed, playing with ChatGPT— not for work, but out of sheer curiosity.&lt;/p&gt;&lt;/blockquote&gt;

&lt;p&gt;A well-written and remarkably transparent note from Josh Miller, CEO of the Browser Company, explaining why the company decided to walk away from Arc, a browser loved by a small but devoted fanbase, in favor of Dia, an AI-first browser. I suspect many are feeling the above in this exact moment -- recognizing that this moment in AI is huge but possessing an allergy to hype cycles.&lt;/p&gt;</content:encoded><source>Letter to Arc members 2025</source></item><item><title>Don&apos;t trust AI to talk accurately about itself</title><link>https://treycausey.com/commonplace/2025-05-26-don-t-trust-ai-to-talk-accurately-about-itself-/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-05-26-don-t-trust-ai-to-talk-accurately-about-itself-/</guid><description>Don&apos;t trust AI to talk accurately about itself</description><pubDate>Mon, 26 May 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;Asking questions like this feels like a natural thing to do: these bots use “I” pronouns (I really wish they wouldn’t) and will very happily answer questions about themselves—what they can do, how they work, even their own opinions (I really wish they wouldn’t do that).&lt;/p&gt;&lt;/blockquote&gt;</content:encoded><source>Don&apos;t trust AI to talk accurately about itself: Bard wasn&apos;t trained on Gmail</source></item><item><title>What AI Thinks It Knows About You</title><link>https://treycausey.com/commonplace/2025-05-25-what-ai-thinks-it-knows-about-you-/</link><guid isPermaLink="true">https://treycausey.com/commonplace/2025-05-25-what-ai-thinks-it-knows-about-you-/</guid><description>Reflections on the opacity of AI models and what they reveal about us</description><pubDate>Sun, 25 May 2025 00:00:00 GMT</pubDate><content:encoded>&lt;blockquote&gt;&lt;p&gt;It calls to mind a maxim about why it is so hard to understand ourselves: &quot;If the human brain were so simple that we could understand it, we would be so simple that we couldn&apos;t.&quot; If models were simple enough for us to grasp what&apos;s going on inside when they run, they&apos;d produce answers so dull that there might not be much payoff to understanding how they came about.&lt;/p&gt;&lt;/blockquote&gt;</content:encoded><source>The Atlantic - What AI Thinks It Knows About You</source></item><item><title>Building Successful Responsible AI Teams</title><link>https://treycausey.com/blog/posts/building_rai_teams/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/building_rai_teams/</guid><description>A practical guide to building responsible AI teams that succeed by aligning with business goals, hiring builders over framework-creators, and focusing on empirical risk management.</description><pubDate>Tue, 19 Nov 2024 00:00:00 GMT</pubDate></item><item><title>Skeptical optimists are the key to an abundant AI future</title><link>https://treycausey.com/blog/posts/skeptical_optimists/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/skeptical_optimists/</guid><description>Beyond doomers and accelerationists: why skeptical optimists who believe AI problems are solvable with effort are essential for building an abundant AI future.</description><pubDate>Tue, 04 Apr 2023 00:00:00 GMT</pubDate></item><item><title>Selecting on the dependent variable</title><link>https://treycausey.com/blog/posts/selecting_dependent_variable/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/selecting_dependent_variable/</guid><description>Why business books and productivity advice often commit the error of selecting on the dependent variable, and how to spot this common reasoning flaw.</description><pubDate>Thu, 03 Nov 2022 00:00:00 GMT</pubDate></item><item><title>Quantification bias</title><link>https://treycausey.com/blog/posts/quantification_bias/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/quantification_bias/</guid><description>Why new systems that afford quantitative measurement face disproportionate scrutiny compared to incumbent systems, and why we should hold all systems to the same standards.</description><pubDate>Sun, 30 Oct 2022 00:00:00 GMT</pubDate></item><item><title>The rise of the data product manager</title><link>https://treycausey.com/blog/posts/rise_of_data_product_manager/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/rise_of_data_product_manager/</guid><description>The emergence of a new kind of product manager who understands data infrastructure, ML, and how to build products where data is at the core of the value proposition.</description><pubDate>Sat, 09 Jul 2022 00:00:00 GMT</pubDate></item><item><title>Software development skills for data scientists</title><link>https://treycausey.com/blog/posts/data_science_interviews/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/data_science_interviews/</guid><description>Notes from my own job search on what to expect in data science interviews, from technical screens to on-site interviews, and how to prepare for both type A and type B data scientist roles.</description><pubDate>Fri, 08 Jul 2022 00:00:00 GMT</pubDate></item><item><title>The acute pain and chronic reward of public-facing work</title><link>https://treycausey.com/blog/posts/emotional_rollercoaster_public_work/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/emotional_rollercoaster_public_work/</guid><description>Reflections on the emotional toll of producing public-facing work like open source or writing, where criticism flows fast while rewards reveal themselves slowly.</description><pubDate>Thu, 07 Jul 2022 00:00:00 GMT</pubDate></item><item><title>Do you have time for a quick chat?</title><link>https://treycausey.com/blog/posts/quick_chat/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/quick_chat/</guid><description>Guidelines for requesting informational coffee chats that respect the other person&apos;s time and increase your chances of getting a response.</description><pubDate>Wed, 06 Jul 2022 00:00:00 GMT</pubDate></item><item><title>Software development skills for data scientists</title><link>https://treycausey.com/blog/posts/software_dev_skills/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/software_dev_skills/</guid><description>Essential software engineering skills that data scientists need to collaborate effectively: version control, code review, testing, documentation, and writing modular code.</description><pubDate>Mon, 04 Jul 2022 00:00:00 GMT</pubDate></item><item><title>Why good data scientists make good product managers</title><link>https://treycausey.com/blog/posts/good_data_scientists_good_pms/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/good_data_scientists_good_pms/</guid><description>The natural affinities between data science and product management, plus the challenges data scientists face when making the transition to PM roles.</description><pubDate>Sun, 03 Jul 2022 00:00:00 GMT</pubDate></item><item><title>Hiring data scientists</title><link>https://treycausey.com/blog/posts/hiring_data_scientists/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/hiring_data_scientists/</guid><description>A better approach to hiring data scientists: treat candidates as intelligent humans, use homework assignments over whiteboard coding, and build a process that leads to equitable outcomes.</description><pubDate>Sat, 02 Jul 2022 00:00:00 GMT</pubDate></item><item><title>Getting started in data science</title><link>https://treycausey.com/blog/posts/getting_started/</link><guid isPermaLink="true">https://treycausey.com/blog/posts/getting_started/</guid><description>A guide to getting started in data science, covering the essential foundations in math, statistics, experiments, machine learning, and tooling that aspiring data scientists need.</description><pubDate>Fri, 01 Jul 2022 00:00:00 GMT</pubDate></item></channel></rss>