<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Doomscrolling Babel]]></title><description><![CDATA[Writing about online platforms and the impact of AI on society.]]></description><link>https://doomscrollingbabel.manoel.xyz</link><image><url>https://substackcdn.com/image/fetch/$s_!ol0M!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1c88597c-d832-45b5-a4ed-5717e2dec675_256x256.png</url><title>Doomscrolling Babel</title><link>https://doomscrollingbabel.manoel.xyz</link></image><generator>Substack</generator><lastBuildDate>Tue, 05 May 2026 11:53:11 GMT</lastBuildDate><atom:link href="https://doomscrollingbabel.manoel.xyz/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Manoel Horta Ribeiro]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[doomscrollingbabel@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[doomscrollingbabel@substack.com]]></itunes:email><itunes:name><![CDATA[Manoel Horta Ribeiro]]></itunes:name></itunes:owner><itunes:author><![CDATA[Manoel Horta Ribeiro]]></itunes:author><googleplay:owner><![CDATA[doomscrollingbabel@substack.com]]></googleplay:owner><googleplay:email><![CDATA[doomscrollingbabel@substack.com]]></googleplay:email><googleplay:author><![CDATA[Manoel Horta Ribeiro]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Qualitative Research After the Interviewer Bot]]></title><description><![CDATA[A new hot take]]></description><link>https://doomscrollingbabel.manoel.xyz/p/qualitative-research-after-the-interviewer</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/qualitative-research-after-the-interviewer</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Sat, 18 Apr 2026 13:16:53 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e89dad8f-a6a3-4c19-a60f-b7dfd0d2d9ff_1085x1080.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p style="text-align: justify;">I have a new hot take. The first half of it is perhaps not so <em>hot</em>, and it would probably be a consensus on Bluesky: <strong>AI cannot replace qualitative research.</strong> As the <a href="https://journals.sagepub.com/doi/full/10.1177/10778004251401851">letter</a> from prominent qualitative researchers argues, meaning-making is a human activity, and even if LLMs could do it, it is unclear how useful it would be. More broadly, I&#8217;d argue that qualitative research is what qualitative researchers do &#8212; and at least for now, they are not buying the AI pill.  The second part is where things may get controversial. <strong>People will use Generative AI to answer questions previously reserved for qualitative research, and they will do so within a quantitative paradigm preferred by institutions. </strong>And before you throw the first stone, this is a descriptive claim, not a normative one.</p><p style="text-align: justify;">But let me take a step back to define what I mean by &#8220;qualitative research,&#8221; because my use of the term is inevitably shaped by my work in HCI and other CS-y fields. The tradition I have in mind includes things like ethnography, participant observation, interviewing, interpretive content analysis, and grounded theory. These are sometimes referred to as &#8220;interpretative&#8221; in contrast to &#8220;positivist&#8221; methods. The acceptance of interpretative methods in tech-adjacent fields was (<a href="https://interactions.acm.org/archive/view/january-february-2024/evaluating-interpretive-research-in-hci?utm_source=chatgpt.com">and still is</a>) uneven and bumpy, but it has happened because of their great utility for making sense of phenomena that do not yield easily to logs or math alone. And it feels like these phenomena are ever-relevant, because important questions seem to be at the edge of what our institutions can quantify: stigma, <a href="https://dl.acm.org/doi/pdf/10.1145/3491102.3502038">domestic violence,</a>  emotional labor.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p style="text-align: justify;">But the mainstream institutional acceptance of qualitative research often rests on the limits of quantitative approaches. Institutions have not truly bought the &#8220;thick-description&#8221; pill. They seek it when quantification is not yet available, credible, or cheap, but are quick to pivot once that is no longer true. We see this pattern across the most diverse areas, from <a href="https://www.iilj.org/publications/measuring-the-world-indicators-human-rights-and-global-governance/">human rights monitoring</a> to <a href="https://academic.oup.com/bjsw/article/45/1/395/1739630">child protection</a>, stakeholders&#8217; preferences for numbers displaced much interpretive work in favor of a couple of metrics and decision trees. And it is worth understanding their viewpoint: they have to navigate complicated trade-offs, and that is arguably easier to do once you have simplified your world into a handful of indicators.</p><p style="text-align: justify;">My point here is not to turn this into a discussion of the <a href="https://en.wikipedia.org/wiki/Economics_imperialism">epistemic validity of different methods</a>, but to give context on what advances in GenAI mean for qualitative research (maybe beyond my modest talents, but hey, this is a blog post). Even if not referred to as &#8220;Qualitative AI research,&#8221; new technologies enabled by GenAI such as open-ended interviewing at scale, or semi-automated inductive coding, enable quantitative approaches to examine questions that were previously too much horse for quant research&#8217;s cowboy. Consider, for example, the recent hype around the <a href="https://www.anthropic.com/research/anthropic-interviewer">&#8220;Anthropic Interviewer.&#8221;</a> The system is no ethnographer in any deep sense, but it can collect reasonably good interviews cheaply and at scale. And that&#8217;s probably preferable for many institutions: they can cheaply get something that fills the same gap as qualitative inquiry but in a form more amenable to scale and quantification.</p><p style="text-align: justify;">One potential criticism of my point is that no questions are inherently &#8220;qualitative.&#8221; Fair enough. But in practice, institutions and firms usually behave as if the distinction exists. They distinguish between questions they see as problems of measurement and optimization, and questions they see as requiring interpretation, context, and meaning-making. A health agency, for instance, would not approach estimating vaccine effectiveness in the same way it approaches understanding why some communities resist vaccination. But what happens when that second task begins to look operationalizable at scale, say, through a Claude-backed interviewer?</p><p style="text-align: justify;">What is most interesting is how hard it is to criticize this approach while staying within an interpretivist frame. There are genuine concerns about AI-conducted interviews. But the critiques most likely to persuade institutional stakeholders often appeal to concepts like reliability, validity, and generalizability, terms that interpretivist traditions have often resisted. This means the threat model here is that AI may reorganize how institutions understand the value of qualitative inquiry. </p><p style="text-align: justify;">As a quantitative researcher myself (who got chastised for being a &#8220;positivist&#8221; only once! I was a master&#8217;s student, give me a break), I find these new developments exciting but at the same time slightly terrifying. These technological advances may help address important questions at scale and with the kinds of guarantees that institutions care about: consistency, comparability, auditability, and speed. But they force harder questions: who gets to decide when an AI-informed account of a phenomenon is &#8220;good enough&#8221;? And whose interests are served when that threshold is set? As <a href="https://www.nature.com/articles/s41586-024-07146-0">Messeri and Crockett</a> remind us, &#8220;increasing productivity does not guarantee an improved understanding of the world.&#8221;</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Brandon Sanderson’s Case Against AI Art]]></title><description><![CDATA[Is anti-AI the new &#8220;back in my day&#8221;?]]></description><link>https://doomscrollingbabel.manoel.xyz/p/brandon-sandersons-case-against-ai</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/brandon-sandersons-case-against-ai</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Sun, 15 Feb 2026 20:22:47 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/6874bc1d-a88f-4298-9a7c-cf0be17ada49_1600x1245.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I&#8217;ve spent a significant chunk of my life reading&#8212;and, more recently, listening to&#8212;fantasy and sci-fi. That&#8217;s probably why I clicked so quickly when YouTube recommended a video of Brandon Sanderson, an author I admire (and who, if I&#8217;m honest, is too productive for me to keep up with), talking through his take on AI art. The video is very good, and I highly recommend you watch Brandon read his essay, even though it contains no chasmfiends or mistwraiths.</p><div id="youtube2-mb3uK-_QkOo" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;mb3uK-_QkOo&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/mb3uK-_QkOo?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p>Brandon&#8217;s essay shows intellectual humility, which I find lacking in the AI Art discourse (a genre that tends to oscillate between victory laps and moral panic). He starts talking about how Roger Ebert has infamously said that &#8220;video games are not art.&#8221; A stance he and his nerdy audience take as <em>prima facie</em> false. He goes on to note that Ebert is no pioneer in calling bullshit on whatever new form of art emerges: earlier iterations included poets dunking on prose writers and portrait artists dunking on photographers. Thus, he wonders: Is he just dunking on the new thing?</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>He ultimately concludes that this is <em>not</em> the case. And the way he gets there is by clearing away the objections that <em>aren&#8217;t</em> doing the real work for him. Yes, he&#8217;s worried about the economics. Yes, he&#8217;s bothered by the ethics of training on artists&#8217; work. Yes, the environmental costs matter. But he insists that even in a world where models are trained only on consented data and run on efficient hardware, something would still feel off.</p><p>The core of his argument is that AI art collapses art into <em>product</em>. Where he argues that art is also (and maybe primarily?) <em>process</em>: the activity through which a person becomes capable of making the thing. He illustrates this with his own early, not-great juvenilia: books that, for him, worked as &#8220;receipts&#8221; of him becoming a writer. Art, thus, is the process happening to the person making it.</p><p>He ends by concluding that this is why his take on AI art is different from Ebert&#8217;s take on video games. Because AI art is fundamentally different from the previous revolutions. And, on a more positive note, since art is what society collectively decides to treat as art, there&#8217;s a simple way of &#8220;winning&#8221; the war against AI art: it suffices that some people decide this form of art is not worth pursuing.</p><p>I&#8217;m sympathetic to the spirit of this&#8212;and still not convinced by the conclusion. I see three problems.</p><p>First, although art is socially constructed, that doesn&#8217;t mean it&#8217;s easily steered. I really wish Brazil still produced music like in the 60s or 70s, and I&#8217;m sure a lot of people do. But our wills are not enough, and that stuff only rarely fuels a banger TikTok. Cultural tastes move due to people&#8217;s will, but also due to what is cheap, frictionless, and easy. So even if we grant Sanderson the premise that art is what we collectively treat as art, it doesn&#8217;t follow that &#8220;collective deciding&#8221; is a crisp process we can just execute.</p><p>Second, I&#8217;m not sure I fully buy the historical analogy. Every revolution looks uniquely alien. When photography emerged, one could claim that photography <em>removed the process</em>, just like AI did. And if photography didn&#8217;t <em>kill</em> painting, it definitely reshaped it. Painting survived by shifting its aims: toward impression, abstraction, expression, and the kinds of seeing that a camera didn&#8217;t naturally deliver.</p><p>Third, Sanderson implicitly argues that there is no process in writing with AI art. But isn&#8217;t writing with AI akin to painting with a camera? If you define &#8220;process&#8221; as the physical act of applying pigment, then yes, photography looks like cheating. But if you define process as the chain of choices that expresses taste and intention, photography is saturated with it. So why, in principle, can&#8217;t AI art be the same? I understand that the slop being sold on Kindle feels subpar, but so did the first photographs when compared to the paintings of their time.</p><p>None of this refutes Sanderson&#8217;s insight: art is more than the artifact&#8212;or as he&#8217;d put it, &#8220;journey before destination.&#8221; But it does complicate his account of why AI art &#8220;feels wrong.&#8221; Maybe the bitter lesson here is that, when the time comes, the new and the strange will always find ways to disappoint us. Still, he need not worry about me: 2026 is the year I finally finish <em>Wind and Truth</em>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Empiricism Gap in Computer Science]]></title><description><![CDATA[So far, I have argued that there is a dissonance between, on the one hand, CS&#8217;s founding myths, curricula, and self-image, and, on the other hand, the modern production of knowledge in computer science.]]></description><link>https://doomscrollingbabel.manoel.xyz/p/the-empiricism-gap-in-computer-science</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/the-empiricism-gap-in-computer-science</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Fri, 26 Dec 2025 11:52:48 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/b95767a8-ca00-477f-9ca3-27b65281ee03_2304x1792.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>So far, <a href="https://doomscrollingbabel.manoel.xyz/p/the-empiricization-of-computer-science">I have argued</a> that there is a dissonance between, on the one hand, CS&#8217;s founding myths, curricula, and self-image, and, on the other hand, the modern production of knowledge in computer science.</p><p>Here, I want to name the dominant style of empiricism that has emerged from that reality, particularly in machine learning. For lack of a better term, I&#8217;ll call it <em>build-and-test empiricism</em>. As the name suggests, this flavor of empiricism treats <em>construction</em> as the primary research method: you build a model, a system, or an algorithm, run it, measure the outcomes, and iterate. It works best when the field shares a common yardstick (benchmarks, leaderboards, standardized testbeds) because then disagreement collapses into a simple question: <em>which approach performs better under the same evaluation?</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>In this <em>modus operandi</em>, knowledge accumulates through artifacts and evidence: a working construction, together with empirical performance results that others can reproduce, compare, and build on. A canonical example is the ImageNet <a href="https://www.image-net.org/challenges/LSVRC/2012/index.php">Large Scale Visual Recognition Challenge (ILSVRC)</a>, which turned image classification into a shared yardstick: everyone trained on the same large dataset and compared on the same evaluation metric. When Krizhevsky, Sutskever, and Hinton entered <a href="https://proceedings.neurips.cc/paper/4824-imagenet-classification-with-deep-convolutional-neural-networks.pdf">AlexNet</a> in ILSVRC 2012, they didn&#8217;t just propose &#8220;a good idea,&#8221; but were able to prove the value of their contribution very easily.</p><p>I&#8217;ll contrast this with another flavor of empiricism, prevalent in political science, economics, and epidemiology. Again, for lack of a better term, I&#8217;ll call it <em>describe-and-defend</em> empiricism. This flavor of empiricism treats inference as the first-class citizen: you formulate a claim about the world, assemble evidence, and argue why the data support that claim. It works best when the field shares a common language for what counts as evidence&#8212;identification, counterfactual reasoning, and uncertainty&#8212;because then disagreement collapses into a more straightforward question: is this claim credible given the design and the assumptions?</p><p>In this <em>modus operandi</em>, knowledge accumulates through the refinement and defense of generalizable statements about reality. A canonical example is the minimum-wage debate. Instead of &#8220;does this model beat that model on ImageNet?&#8221;, the shared question became a claim about the world: do minimum-wage increases reduce employment? <a href="https://davidcard.berkeley.edu/papers/njmin-aer.pd">Card and Krueger (1994)</a>&#8217;s famous New Jersey&#8211;Pennsylvania study framed the problem as a quasi-experiment: New Jersey raised its minimum wage in 1992, Pennsylvania did not, so that you can compare employment changes in fast-food restaurants across the border before and after the policy change (a classic difference-in-differences design).  Their headline result challenged the textbook prediction, sparking years of dispute not only about the answer but also about whether the identification strategy, measurement, and data sources justified the claim.</p><p>That is describe-and-defend empiricism in action: the conversation quickly moved to critiques, reanalyses, and replies (e.g., <a href="https://www.nber.org/system/files/working_papers/w12663/w12663.pdf">Neumark &amp; Wascher&#8217;s comment</a> and <a href="https://www.aeaweb.org/articles?id=10.1257/aer.90.5.1397">Card &amp; Krueger&#8217;s reply</a> in the <em>AER</em>), because the community is built around adversarial scrutiny of credibility.  The debate then generalized to broader designs, such as using discontinuities at state borders across multiple policy changes, to test whether the claim holds up across settings.</p><h1>Maintenance Layer</h1><p>A helpful way to contrast these different flavors of empiricism more sharply is to separate the <em>engine</em> of each style from its <em>maintenance layer. </em>In other words, once a community has a means of generating results, how does it ensure honesty? What are the routines and institutions that notice when the engine starts producing nonsense: when results don&#8217;t replicate, when benchmarks get gamed, when claims are overstated, or when an entire methodological approach is proven incorrect? The maintenance layer is the answer: the field&#8217;s error-correction machinery, the set of practices that turns &#8220;a pile of results&#8221; into something closer to cumulative knowledge.</p><p>As we&#8217;ve mentioned, build-and-test empiricism has an apparent engine: artifacts plus shared yardsticks. But shared yardsticks come with predictable pathologies. Because benchmarks are finite and public, it is easy to overfit not only models to data but scientists to metrics: countless small choices (hyperparameters, prompt formats, data filtering, selection of seeds) get tuned until they look good on this evaluation. Reproducibility also becomes fragile: in modern ML, results can turn on details that are hard to fully specify&#8212;hardware, distributed training quirks, nondeterminism, dataset versions, and preprocessing pipelines. And even when a paper&#8217;s headline number is &#8220;real,&#8221; it can be a narrow kind of reality: a win that evaporates under distribution shift, under a slightly different label definition, or in a deployment setting that wasn&#8217;t represented in the benchmark.</p><p>The community&#8217;s response is often to patch the process with <em>rituals</em>&#8212;lightweight forms of quality control meant to raise the floor without slowing the iteration loop too much. Some rituals target reporting (checklists; standard tables and ablations), some target auditability (artifact evaluation and badging; code and data release norms), and some target interpretability and scope (model cards, datasheets, documentation templates). None of these rituals &#8220;solves&#8221; the epistemology of generalization. But they are a pragmatic maintenance strategy: a way to keep the engine running while limiting the most common failure modes, and to make it harder for the field to confuse a brittle benchmark win with a robust piece of knowledge.</p><p>Describe-and-defend empiricism has a different engine: claims anchored in an argument about why the data identify what you say they identify. Its maintenance layer looks different, too. Because the risks are not &#8220;overfitting the benchmark&#8221; so much as &#8220;fooling ourselves with confounding, researcher discretion, or fragile specifications.&#8221; Thus, the field expends considerable effort on robustness: alternative specifications, placebos, sensitivity analyses, replication across contexts, and explicit uncertainty quantification. The rituals exist here too, pre-analysis plans, preregistration, replication packages, but they are downstream of a deeper norm: you do not get to make a strong claim unless you can defend it.</p><p>No better way to see the difference between the two <em>maintenance layers</em> than to observe how each ecosystem responds when it discovers a methodological error. In econometrics, the recent &#8220;difference-in-differences reckoning&#8221; is a good example: researchers realized that the vanilla two-way fixed effects DiD regression behaves badly under staggered adoption and heterogeneous treatment effects <a href="https://www.sciencedirect.com/science/article/abs/pii/S0304407621001445">(Goodman-Bacon, 2021)</a>. In plain English, people were doing bad stats for a couple of years. The response was not to &#8220;use a checklist next time.&#8221; Instead, the field produced <a href="https://www.sciencedirect.com/science/article/pii/S0304407620303948">new estimators</a>, <a href="https://www.sciencedirect.com/science/article/pii/S0304407623001318">review papers with concrete recommendations</a>, and, crucially, <em>a wave of re-analyses</em>: applied papers were re-estimated with updated methods to see which conclusions survived (e.g., <a href="https://www.aeaweb.org/articles?id=10.1257/app.20230089">Nagengast and Yotov, 2025</a>).</p><p>Machine learning handled &#8220;mistakes&#8221; differently. For example, when a benchmark is found to be leaky, saturated, or overly gameable, the community typically shifts the goalposts rather than revisiting the back catalog. Consider ImageNet: when <a href="https://arxiv.org/abs/1902.10811">Recht et al. (2019)</a> built <em>fresh</em> test sets and found nontrivial accuracy drops, evidence that the community had partially overfit to a heavily reused benchmark. The typical response was not to issue corrigenda for years of ImageNet papers. Instead, the field continued to iterate on the next yardstick. E.g., by the time Recht published his paper, computer vision&#8217;s center of gravity had already shifted from image classification to tasks such as detection, segmentation, and captioning, and to datasets such as <a href="https://arxiv.org/abs/2403.18819">MS COCO</a>. Old papers remain &#8220;fine,&#8221; not because their conclusions are timeless, but because they were produced under the rituals that were locally acceptable at the time. The maintenance layer is oriented toward &#8220;patch-and-proceed&#8221; rather than &#8220;re-estimate-and-reckon&#8221;.</p><p><strong>Why build-and-test empiricism is not enough</strong></p><p>This piece is a critique of building-and-testing empiricism. But it is worth steelmanning it before criticizing. In many ways, the progress in ML and CS more generally is the envy of other disciplines. Build-and-test empiricism creates fast feedback loops, rewards concrete contributions, and makes knowledge unusually transmissible. Under that regime, research &#8220;compounds&#8221;: the community shares datasets and code, converges on testbeds, and iterates quickly enough to discover regimes that theory did not predict. Moritz Hardt put this better than I ever could in his <a href="https://mlbenchmarks.org/">new book on ML benchmarks</a>: <em>&#8220;researchers might&#8217;ve indeed dismissed the idea of benchmarking as unreasonable, if it hadn&#8217;t been so successful.&#8221;</em></p><p>So what&#8217;s the problem? Especially outside of machine learning, CS increasingly studies domains that lack a benchmark to guide iterative development, and many of the most critical questions are explicitly causal: Does this interface change behavior? Does this policy reduce harm? Does this intervention improve well-being? When build-and-test empiricism loses its yardstick and lacks a shared causal language, it becomes easy to produce papers that look empirical but do not cumulate. It is unclear what claim was actually identified, and unclear what a &#8220;better&#8221; result should even mean.</p><p>The deeper issue is that we often keep a build-and-test mentality even when the object of study has shifted from artifacts to claims. In ML, a narrow win can be &#8220;fine&#8221; because the claim is narrow: this method improves this metric under this protocol. But in HCI, privacy, and policy-adjacent CS, the implied claim is usually broader: that some design choice will change how people behave, or that an intervention will improve real outcomes. Those are high-stakes claims&#8212;and yet our field rarely treats them with the institutional seriousness that claim-centered disciplines have built up over decades: routine re-analyses when a standard method is criticized, explicit norms for what constitutes robustness, or a culture in which the community revisits and revises empirical conclusions.</p><p>This gap is most evident in the genre expectations of our papers. The typical &#8220;implications&#8221; section in CHI is written as if it were a policy memo, even when the evidentiary base is modest: a lab study with a convenience sample, a deployment in a single setting, or interviews with a dozen participants. None of these is an invalid method. But they support a particular kind of claim, under particular scope conditions. When we glide from &#8220;here is what we observed in this context&#8221; to &#8220;here is how platforms serving millions should behave,&#8221; we invite exactly the kind of scrutiny that economics seminars institutionalize: what is the counterfactual, what is the estimand, what is the uncertainty, and what would change your conclusion? This mismatch between our ambition and the discipline of our inferential machinery undermines CS&#8217;s ability to shape the real world and weakens its credibility with policymakers and other audiences trained to ask such questions.</p><p>But the hole is even deeper, and there are now problems even in the most &#8220;yardstickable&#8221; subfields. The trouble is that the same properties that make this style powerful also make it easy to confuse local wins for general truths. A benchmark score indicates that a method performed well under a specific evaluation protocol. When the field treats &#8220;state of the art&#8221; as synonymous with &#8220;we now understand the phenomenon,&#8221; it blurs the distinction between works here and works in general, and between what works and what is true. Worse, the benchmark can quietly become the target: we optimize for the metric, and then retrofit narratives about the underlying capability.</p><p>If you want examples here, look no further than the COVID craze, the few months where researchers from every field (including myself) abandoned whatever they were doing to do &#8220;a COVID-related&#8221; paper. In ML, the result was a flood of models with glossy AUCs and clean-looking tables, but surprisingly little durable knowledge. <a href="https://www.nature.com/articles/s42256-021-00307-0">Roberts et al. (2021)</a> argued that much of the literature suffered from predictable methodological pitfalls&#8212;dataset bias, leakage, non-representative sampling, and weak evaluation protocols. These were severe enough that, in their study, none of the 2,212 models proposed to detect and prognosticate COVID-19 from chest radiographs and CT scans were considered potentially clinically useful. Take a second to consider this. It is just insane. 2,212 papers  were misled by creating their own &#8220;local&#8221; win through a benchmark and ignoring the real problem they were supposed to be solving.</p><p>In this case, the &#8220;yardstick&#8221; is still somewhat crisp, and you could imagine a responsible researcher engaging more seriously with the literature and getting something worthwhile. Yet, machine learning and computer science, more generally, are increasingly encroaching on domains where the &#8220;yardstick&#8221; is no longer crisp and where there is not a single yardstick, but multiple. Generative AI has dragged the field toward tasks that are more abstract, open-ended, and value-laden: helpfulness, harmlessness, truthfulness, creativity, persuasion, alignment, &#8220;good judgment.&#8221;</p><p>As a result, the build-and-test engine begins to wobble: we still want shared yardsticks, but the target is moving, and the metric often serves as a proxy for something we cannot directly observe. The field responds in its characteristic way, by building new benchmarks, rubrics, and evaluator models.  Still, under this new regimen, it is much easier for &#8220;what scores well&#8221; to drift away from &#8220;what is actually good,&#8221; and much harder to know whether an apparent improvement reflects a real capability gain or simply better exploitation of the measurement procedure.</p><p><strong>Why can&#8217;t we do both?</strong></p><p>This is where the describe-and-defend approach to empiricism offers an essential lesson for CS. It provides a shared language for what a claim <em>entails: its assumptions, counterfactuals, uncertainty, and </em>scope. Without that language, the community tends to reach for rituals as a substitute for reasoning: more checklists, more templates, more badges. These rituals often help. But they are a thin proxy for the greater skill of asking: what exactly does this result establish, and what would it take for it to stop being true?</p><p>A recent position paper by <a href="https://arxiv.org/abs/2502.00561?">Wallach et al. (2025)</a> makes this point, but aimed squarely at the GenAI moment: evaluating generative AI systems is a social science measurement challenge. They argue that many GenAI evaluations amount to &#8220;apples-to-oranges comparisons,&#8221; and that the ML community would benefit from importing measurement theory: being explicit about <em>what</em> construct you claim to measure (capabilities, behaviors, impacts), how your instrument operationalizes it, and what validity evidence would justify treating the result as knowledge rather than noise. \</p><p>But before we get too trigger-happy about the conclusion that &#8220;we should all be social scientists now,&#8221; it is worth scrutinizing describe-and-defend empiricism for the messy beast it also is. For one, it is often slow. When the unit of progress is a defensible claim about the world, you pay for credibility in calendar time: collecting better data, negotiating access, validating measures, and stress-testing identification strategies. And even then, the best you can sometimes do is a careful estimate under plausible assumptions, but never fully provable.</p><p>Worse, describe-and-defend carries its own failure modes. When the reward is a clean claim, it can incentivize overconfident narratives, specification search, and knife-edge identification strategies that look airtight on paper but are fragile in practice. Robustness checks help, but they can also become rituals of their own, performed because reviewers expect them, not because they genuinely change anyone&#8217;s mind. Even credible causal estimates can be narrow: a precisely defended effect in one setting that is difficult to generalize, interpret mechanistically, or translate into design guidance.</p><p>The punchline is not &#8220;CS should become economics&#8221; or &#8220;every paper needs a preregistration.&#8221; The punchline is that CS needs to become bilingual. We should keep build-and-test empiricism&#8212;its speed, openness, and compounding progress are real achievements&#8212;but we should supplement it with a second language of empiricism that lets us state and evaluate claims with more precision. That language helps us separate exploration from confirmation, local wins from generalization, and performance improvements from causal explanation. In other words: keep the engine, but upgrade the epistemology.</p><p>What would this look like in practice? <a href="https://doomscrollingbabel.manoel.xyz/p/the-missing-discipline-in-computer">It starts with teaching a small set of concepts that claim-centered fields treat as basic literacy</a>: what is the estimand we are trying to learn, what is the counterfactual, what assumptions connect the data we have to the claim we want, what are the threats to identification, what uncertainty do we actually have, and what robustness checks would meaningfully stress the conclusion. It also means taking measurement more seriously in settings where CS is rapidly expanding (especially in generative AI) by being explicit about constructs, validity, and scope.</p><p>Finally, it requires a cultural shift in the rewards we value. Not every paper needs to make a sweeping claim; sometimes the honest contribution is a careful measurement instrument, a negative result, or a boundary condition that shows where an approach breaks. And when a methodology is shown to be flawed, we should normalize reanalysis as part of the scientific process. If build-and-test is how CS moves fast, then principled inference is how it learns what it has actually learned.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Empiricization of Computer Science]]></title><description><![CDATA[Scientific disciplines, like nations, have their own founding myths.]]></description><link>https://doomscrollingbabel.manoel.xyz/p/the-empiricization-of-computer-science</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/the-empiricization-of-computer-science</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Wed, 17 Dec 2025 13:28:31 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/8f1d7160-1bfc-4c1f-9ba8-7fcf0641b351_2304x1792.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Scientific disciplines, like nations, have their own founding myths. These may sound inconsequential at first, but they deeply shape how disciplines see themselves, what they value, and what they believe their role in the world ought to be. I&#8217;d argue that computer science was built upon founding myths around theory and system building. Princeton has a very exciting course called <a href="https://registrar.princeton.edu/course-offerings/course-details?term=1264&amp;courseid=010571">&#8220;Great Moments of Computing,&#8221;</a> where Margaret Martonosi walks students from the foundations of digital logic and differential privacy to the creation of network protocols and the computer mouse. In a sense, these are the closest equivalents computer scientists have to a Durkheim, a Smith, or a Freud.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yyYn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ecfd2bf-5e73-4b56-a566-2da5d7f73015_770x354.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yyYn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ecfd2bf-5e73-4b56-a566-2da5d7f73015_770x354.png 424w, https://substackcdn.com/image/fetch/$s_!yyYn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ecfd2bf-5e73-4b56-a566-2da5d7f73015_770x354.png 848w, https://substackcdn.com/image/fetch/$s_!yyYn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ecfd2bf-5e73-4b56-a566-2da5d7f73015_770x354.png 1272w, https://substackcdn.com/image/fetch/$s_!yyYn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ecfd2bf-5e73-4b56-a566-2da5d7f73015_770x354.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yyYn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ecfd2bf-5e73-4b56-a566-2da5d7f73015_770x354.png" width="478" height="219.75584415584416" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0ecfd2bf-5e73-4b56-a566-2da5d7f73015_770x354.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:354,&quot;width&quot;:770,&quot;resizeWidth&quot;:478,&quot;bytes&quot;:56175,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://doomscrollingbabel.manoel.xyz/i/181885969?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ecfd2bf-5e73-4b56-a566-2da5d7f73015_770x354.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yyYn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ecfd2bf-5e73-4b56-a566-2da5d7f73015_770x354.png 424w, https://substackcdn.com/image/fetch/$s_!yyYn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ecfd2bf-5e73-4b56-a566-2da5d7f73015_770x354.png 848w, https://substackcdn.com/image/fetch/$s_!yyYn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ecfd2bf-5e73-4b56-a566-2da5d7f73015_770x354.png 1272w, https://substackcdn.com/image/fetch/$s_!yyYn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ecfd2bf-5e73-4b56-a566-2da5d7f73015_770x354.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a><figcaption class="image-caption">Examples of the moments covered in Martonosi's course.</figcaption></figure></div><p>Yet the nature of computer science has changed tremendously: it is now an empirical science, driven by observation and experiment as much as by theory and construction. If you don&#8217;t believe me, spend 5 minutes going over the best paper awards for three conferences in different subfields in CS. If you did so in 2025, you might have found papers like <a href="https://www.usenix.org/conference/usenixsecurity25/presentation/kireev">&#8220;Characterizing and Detecting Propaganda-Spreading Accounts on Telegram&#8221;</a> (USENIX; Security and Privacy), or <a href="https://dl.acm.org/doi/10.1145/3757537">&#8220;Examining Mental Health Conversations with Large Language Models through Reddit Analysis&#8221;</a> (CSCW; Human Computer Interaction), or <a href="https://openreview.net/forum?id=s0JVsx3bx1">&#8220;Scaling Depth Can Enable New Goal-Reaching Capabilities&#8221;</a> (NeuRIPS; Machine Learning). The first two papers describe online user traces, to understand sociotechnical phenomena (Online Propaganda, people using LLMs to talk about mental health issues), while the latter is a full-blown examination of what happens when you do self-supervised RL with very deep neural networks.</p><p>This empiricization of Computer Science seems to have happened independently across subdisciplines, and I argue that the reasons are two-fold.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2><strong>Reason #1: Socio-technical integration of computing</strong></h2><p>First, computer scientists steadily broadened the scope of what they treated as &#8220;a computing problem.&#8221; Part of this is simply baked into the field&#8217;s origins: computing inherited deep roots in mathematics and engineering, but it also grew up in constant contact with other disciplines: think <a href="https://en.wikipedia.org/wiki/Cybernetics">cybernetics</a> or <a href="https://en.wikipedia.org/wiki/Cognitive_science">cognitive science</a> As computing technologies matured, they also began to fade into the background of everyday life, becoming infrastructure rather than an object of conscious attention. In his seminal <a href="https://dl.acm.org/doi/10.1145/329124.329126">&#8220;The Computer for the 21st Century&#8221;</a> essay Mark Weiser captured the consequence of that transition: <em>&#8220;only when things disappear in this way are we freed to use them without thinking and so to focus beyond them on new goals.&#8221;</em></p><p>Privacy is a particularly vivid example. In an <a href="https://www.nationalacademies.org/read/12027/chapter/4#6">earlier framing,</a> &#8220;privacy&#8221; looks like a technical property of systems&#8212;confidentiality, secure channels, access control&#8212;problems that naturally invite cryptographic solutions. But as privacy disputes increasingly played out in real social settings, the field had to confront privacy as something defined by <a href="https://dl.acm.org/doi/10.1145/1242572.1242661">people&#8217;s expectations, norms, and context</a>. <a href="https://heinonline.org/hol-cgi-bin/get_pdf.cgi?handle=hein.journals/washlr79&amp;section=16&amp;casa_token=SzIDbe_3VDMAAAAA:mVwgHHhHOEbB2odfGbxAP6Nyp2tC_BP5hbWGmPYvgX9tW465FjIk7NngciqRcaj9Vx7j6h4Z">Nissenbaum&#8217;s theory of contextual integrity</a> crystallizes this shift by tying privacy to whether information flows are appropriate relative to contextual norms. Even when the cryptography is excellent, privacy can still fail at the human interface: if users pick weak or reused passwords, attackers don&#8217;t need to &#8220;break&#8221; encryption, just to log in!</p><p>Another example is the evolution&#8212;and very existence&#8212;of the field of human&#8211;computer interaction. <a href="https://dl.acm.org/doi/book/10.1145/2594128">SIGCHI&#8217;s early curricular definition</a> makes the scope expansion explicit: HCI is concerned with the design, evaluation, and implementation of interactive systems &#8220;for human use,&#8221; and with &#8220;the study of major phenomena surrounding them.&#8221; In other words, once the object of study is interaction rather than computation alone, empiricism is no longer optional: you need user studies, experiments, field observations, and iterative evaluation to know whether a system is usable, safe, or meaningful. Over time, HCI also widened what it counts as &#8220;interaction&#8221;. <a href="https://dl.acm.org/doi/abs/10.1145/1182475.1182476">B&#248;dker (2006)</a> describes the field as moving from a first wave rooted in human factors and design, to a second wave shaped by cognitive science, and then to a third wave emphasizing situated meaning, values, and social context.</p><p>Just as HCI institutionalized user studies, the systems and networking communities institutionalized measurement itself. SIGMETRICS <a href="https://www.sigmetrics.org/BriefHistoryofSIGMETRICS.pdf">emerged as a way to measure the differences in performance across computers</a> (a subtly challenging task). And while you could argue there&#8217;s no &#8220;societal factor&#8221; here, this changed the moment the Internet became a live, evolving, operational system. Core properties of the Web &#8220;worked&#8221; were elusive because they relied on how people and institutions used it in practice. This led to the need for a new, much more empirical type of research that required sophisticated measurement infrastructure. For instance the <a href="https://conferences.sigcomm.org/imc/2001/cfp.html">SIGCOMM Internet Measurement Workshop (2001)</a> explicitly asked for work that improves understanding of how to collect/analyze measurements or &#8220;give insight into how the Internet behaves,&#8221; and it stated bluntly that papers not relying on measurement were &#8220;out of scope.&#8221;</p><p>Finally, the rise of computational social science blurs the boundary between &#8220;CS papers&#8221; and &#8220;social science papers.&#8221; Lazer and colleagues open <a href="https://www.science.org/doi/10.1126/science.1167742">their manifesto</a> with the observation that &#8220;we live life in the network,&#8221; leaving digital traces that can be assembled into detailed pictures of individual and collective behavior. In that world, it becomes increasingly hard to draw a clean line between disciplines: the same project may involve building data pipelines, designing machine learning models, running quasi-experiments and interpreting results through social science theories. Go to a venue like IC2S2, and you&#8217;ll find people in CS working on exactly the same kinds of problems as people in Sociology or Political Science.</p><p>Taken together, these shifts mark a quiet redefinition of the discipline: as computing &#8220;disappeared&#8221; into infrastructure, the unit of analysis stopped being the machine in isolation and became the human&#8211;system&#8211;institution loop.</p><h2><strong>Reason #2: The Triumph of Empiricism</strong></h2><p>Second, you have the triumph of empiricism, particularly in machine learning. If you dial back to the late 90s and early 00s, neural network, or &#8220;connectionism,&#8221; was remarkably unfashionable. In a <a href="https://spectrum.ieee.org/facebook-ai-director-yann-lecun-on-deep-learning">2015 interview with IEEE Spectrum</a>, Turing Awardee Yann LeCun said: <em>&#8220;I have very little patience for people who do theory about a particular thing simply because it&#8217;s easy&#8212;particularly if they dismiss other methods that actually work empirically, just because the theory is too difficult. There is a bit of that in the machine learning community. In fact, to some extent, the &#8220;Neural Net Winter&#8221; during the late 1990s and early 2000s was a consequence of that philosophy; that you had to have ironclad theory, and the empirical results didn&#8217;t count.&#8221;</em></p><p>But at the same time as neural networks were &#8220;shunned,&#8221; something else was happening in Machine Learning Land: the discipline started to become increasingly oriented around shared tasks and competitions. Perhaps this was only natural for, as one of our founding myths is the Turing test: a challenge were a machine must exhibit behavior indistinguishable to that of a human. Isabelle Guyon, a co-inventor of the support vector machine, shares a brief history of data in her <a href="https://nips.cc/virtual/2022/invited-talk/56158">2022 NeurIPS talk</a> titled &#8220;How ML is Becoming an Experimental Science&#8221;. She notes that the field evolved from toy data, to shared datasets (most prominently made available in the UCI ML Repository) to larger datasets that were often embedded in a explicitly competitive context. Some prominent early competitions include the KDD Cup, hosted by the Association for Computing Machinery (ACM), and competitions hosted by the National Institute of Standards and Technology (NIST), one of which introduced the now famous <a href="https://en.wikipedia.org/wiki/MNIST_database">MNIST digits database</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BxKo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7cacf8-9d73-44a3-b14a-560642b2869a_1872x642.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BxKo!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7cacf8-9d73-44a3-b14a-560642b2869a_1872x642.png 424w, https://substackcdn.com/image/fetch/$s_!BxKo!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7cacf8-9d73-44a3-b14a-560642b2869a_1872x642.png 848w, https://substackcdn.com/image/fetch/$s_!BxKo!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7cacf8-9d73-44a3-b14a-560642b2869a_1872x642.png 1272w, https://substackcdn.com/image/fetch/$s_!BxKo!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7cacf8-9d73-44a3-b14a-560642b2869a_1872x642.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BxKo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7cacf8-9d73-44a3-b14a-560642b2869a_1872x642.png" width="1456" height="499" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1a7cacf8-9d73-44a3-b14a-560642b2869a_1872x642.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:499,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1658886,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://doomscrollingbabel.manoel.xyz/i/181885969?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7cacf8-9d73-44a3-b14a-560642b2869a_1872x642.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BxKo!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7cacf8-9d73-44a3-b14a-560642b2869a_1872x642.png 424w, https://substackcdn.com/image/fetch/$s_!BxKo!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7cacf8-9d73-44a3-b14a-560642b2869a_1872x642.png 848w, https://substackcdn.com/image/fetch/$s_!BxKo!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7cacf8-9d73-44a3-b14a-560642b2869a_1872x642.png 1272w, https://substackcdn.com/image/fetch/$s_!BxKo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a7cacf8-9d73-44a3-b14a-560642b2869a_1872x642.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Mandatory ImageNet image.</figcaption></figure></div><p>As compute and data became plentiful, deep learning flourished. A canonical turning point is AlexNet&#8217;s win at the <a href="https://www.image-net.org/challenges/LSVRC/2012/">2012 ImageNet Large Scale Visual Recognition Challenge (ILSVRC).</a> ImageNet was unusually ambitious for its time&#8212;an effort to scale visual recognition with millions of labeled images. And ILSVRC distilled that vision into a standardized benchmark of 1,000 categories with about 1.2M training images. In 2012, <a href="https://proceedings.neurips.cc/paper_files/paper/2012/file/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf">Krizhevsky, Sutskever, and Hinton</a> trained a large convolutional network on this dataset and crushed the field: their entry achieved a 15.3% top-5 error, compared to 26.2% for the runner-up.</p><p>The increasing adoption of doing &#8220;what works&#8221; marked an inversion of how knowledge ought to be accumulated in machine learning. Before, experiments were seen as confirmatory. The idea was that better theory would lead to better models, and that experiments were a way to confirm that a model was indeed better (and that perhaps the theory did not have any flaws). For example, in a <a href="https://ieeexplore.ieee.org/document/788640">1999 piece</a>, Vapnik describes the rise of support vector machines in the mid-1990s as learning algorithms &#8220;based on the developed theory,&#8221; explicitly positioning statistical learning theory not only as analysis but as a driver of practical algorithm design.</p><p>In the deep learning era, the direction often runs the other way: models are frequently guided by intuition and engineering heuristics first, with theoretical framing arriving later (if at all). The original dropout paper is almost playful about this, stating that the motivation for the method &#8220;comes from a theory of the role of sex in evolution (...). A closely related motivation for dropout comes from thinking about successful conspiracies&#8221; <a href="https://jmlr.org/papers/v15/srivastava14a.html">(Srivastava, 2014)</a>.</p><p>Theory, then, often enters as post hoc explanation and consolidation. A prominent follow-up to dropout was literally titled <a href="https://proceedings.neurips.cc/paper_files/paper/2013/file/71f6278d140af599e06ad9bf1ba03cb0-Paper.pdf_">&#8220;Understanding Dropout,&#8221;</a>, and it set out to formalize what dropout is doing (as averaging/regularization, and in simple settings as optimizing a regularized objective), giving a theoretical account for an empirically successful trick. A similar arc shows up in the &#8220;double descent&#8221; story: modern overparameterized models displayed test-error behavior that looked inconsistent with the textbook bias&#8211;variance tradeoff, and only afterward did theory work emerge to reconcile the classical U-shape with the empirical &#8220;interpolation peak&#8221; and the subsequent performance improvement at larger scales. Again, theory did not inform that models should be overparameterized so that at some point, the error rate would go down again. People just tried doing things and some of them worked well.</p><p>And someone might argue: &#8220;But this is only machine learning&#8212;CS is so broad!&#8221; But is it really? The meteoric rise of machine learning&#8212;and, more recently, generative AI&#8212;has started to behave less like one subfield and more like a method layer that percolates through the rest of computer science (and increasingly, through the sciences more broadly). In 2024, the Physics Nobel prize went to Hopfield and Hinton &#8220;for foundational discoveries and inventions that enable machine learning with artificial neural networks,&#8221; and the Chemistry prize went (in part) to Hassabis and Jumper for an AI model that solved the protein-structure prediction problem (AlphaFold). Mathematics may be at the cusp of a revolution in which language models become integral to cutting-edge proofs&#8212;<a href="https://terrytao.wordpress.com/2025/12/08/the-story-of-erdos-problem-126/">as Terence Tao&#8217;s recent account of the human&#8211;AI collaboration suggests</a>. On the CS side, conferences like MLSys exist precisely because &#8220;systems&#8221; and &#8220;machine learning&#8221; have become deeply entangled&#8212;enough to warrant a dedicated flagship venue at their intersection. And in HCI, a growing slice of the field is now about <a href="https://dl.acm.org/doi/abs/10.1145/3630106.3658941">designing with the inherent messiness of language models</a>, leading to new interface patterns and evaluations aimed at helping users cope with these failure modes.</p><h2><strong>So what?</strong></h2><p>This whole rant started as a &#8220;we don&#8217;t have empiricism as a founding myth.&#8221; But then, I&#8217;ve spent quite a bit of ink arguing <em>that</em> the empiricization of computer science has already happened. So what really is the problem here? </p><p>The problem is <em>how</em> we became empirical. Too often, we practice a kind of post-hoc, radical empiricism: run what we can run, keep what works, and only afterward scramble for the vocabulary of validity, uncertainty, and explanation. The tell is that we keep patching rigor in <em>after the fact</em> (through checklists, artifact badging, and community process fixes) because we never made principled empiricism a first-class part of the curriculum. </p><p>Watch for the next post for more on that and on what a more principled empiricism would entail in CS.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Labeling Data with Language Models: Trick or Treat? ]]></title><description><![CDATA[Large language models are now labeling data for us.]]></description><link>https://doomscrollingbabel.manoel.xyz/p/labeling-data-with-language-models</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/labeling-data-with-language-models</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Sat, 25 Oct 2025 12:13:18 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/5446ec7b-3cb0-4f3a-a61b-e9e68e467e8f_2304x1792.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Large language models are now labeling data for us. In dozens of new studies, researchers have quietly replaced undergrad coders and Mechanical Turkers with GPT4 or Gemini 2.5. What began as a convenience is fast becoming a methodological revolution! And one that could reshape how social scientists measure the world.</p><p>As I&#8217;ve been catching up with the literature, I see three concurrent &#8220;waves&#8221; of work around this idea.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Early papers were awed by LMs&#8217; annotating abilities. Back in 2023, <a href="https://www.pnas.org/doi/abs/10.1073/pnas.2305016120">Gillardi et al.</a> showed that &#8220;ChatGPT outperforms crowd workers for several annotation tasks, including relevance, stance, topics, and frame detection.&#8221; And it did so at a fraction of the cost&#8212;roughly thirty times cheaper than human annotators. Similar results quickly followed, suggesting that LMs might transform quantitative social science (see <a href="https://journals.sagepub.com/doi/full/10.1177/08944393241286471">T&#246;rnberg</a>, or <a href="https://x.com/ryancbriggs/status/1980370365743534455">this recent tweet</a>). In ML and NLP circles, this trend became known as <em>LM-as-a-judge</em> (<a href="https://proceedings.neurips.cc/paper_files/paper/2023/hash/91f18a1287b398d378ef22505bf41832-Abstract-Datasets_and_Benchmarks.html">Zheng et al.</a>; <a href="https://openreview.net/forum?id=8euJaTveKw">Kim et al.</a>)</p><p>The next wave asked a harder question: <em>how do we do this right?</em> <a href="https://direct.mit.edu/coli/article/50/1/237/118498">Ziems et al.</a> systematically analyzed LM annotations across models and tasks, producing general usage guidelines. Others accepted that LM labels will never be perfect and proposed ways to account for their bias. Two go-to methods here are DSL by <a href="https://dl.acm.org/doi/10.5555/3666122.3669122">Egami et al.</a> and CDI by <a href="https://aclanthology.org/2025.naacl-long.179/">Gligori&#263; et al.</a> The basic idea of these approaches is to use a smaller set of human annotations to &#8220;ground&#8221; the LM-generated labels. Once you model the LM bias relative to the human sample, you can then correct your final estimates. <a href="https://aclanthology.org/2025.acl-long.782/">Calderon et al.</a> go a slightly different route. They use a sample of data annotated multiple times in an &#8220;alternative annotator test,&#8221; which justifies using an LM rather than a human annotator (because it outperforms annotators in aligning with the majority vote).</p><p>Finally, a &#8220;third wave&#8221; of work has currently raised the alarm that LMs-as-annotators might lead to bad science. In perhaps my favorite paper of this wave, <a href="https://arthurspirling.org/documents/BarriePalmerSpirling_TrustMeBro.pdf">Barrie et al.</a> show that even controlling the temperature, using proprietary models as annotators leads to unacceptably high variance in performance over the course of many months. <a href="https://x.com/ey_985/status/1980272634458951998">Yang et al.</a> replicated 14 recently published papers to find that &#8220;<em>LLM annotations have low intercoder reliability with the original annotations and moderate reliability among the LLMs themselves.&#8221; </em><a href="https://arxiv.org/abs/2509.08825">Baumann et al.</a> find that conclusions drawn from LLM annotations often differ from those from human annotations, depending on the LLM used and its configuration. They define this term &#8220;LM hacking&#8221; as mistakes that &#8220;propagate to downstream analyses and cause Type I (false positive), Type II (false negative), Type S (wrong sign), or Type M (exaggerated effect) errors.&#8221;</p><p>Okay, so where do we go from here?</p><p>First, we need to clarify what problem we&#8217;re actually solving.</p><p>For example, the aforementioned LM hacking paper by <a href="https://arxiv.org/abs/2509.08825">Baumann et al.</a> groups all errors under the same umbrella, e.g., they consider not finding an effect when one exists as &#8220;LM hacking.&#8221; The paper then concludes that methods such as DSL and CDI are not &#8220;enough&#8221; to address LM hacking. This conflates <em>error type</em> with <em>scientific risk</em>. DSL and CDI explicitly treat LM labels as noisy measurements: they use a small human &#8220;gold&#8221; set to estimate measurement error and then debias or propagate that uncertainty. By design, they trade some power (more Type II) to reduce false discovery (Type I)&#8212;the same logic behind strict multiple-comparison controls (e.g., Bonferroni increases Type II while lowering FWER)</p><p>So, really, the meta-science question here is about efficiency, not purity. In the social sciences (and perhaps in general), we are currently constrained to explore only a very small subset of the total space of interesting questions worth answering (<a href="https://www.cambridge.org/core/journals/behavioral-and-brain-sciences/article/beyond-playing-20-questions-with-nature-integrative-experiment-design-in-the-social-and-behavioral-sciences/7E0D34D5AE2EFB9C0902414C23E0C292">see this banger</a>; <a href="https://drive.google.com/file/d/1-4kcD8yn4dTikxrbbm5oaGnaEvj3XQ5B/view">or this one</a>). So let's ask: <em>how can we produce scientific knowledge more efficiently?</em></p><p>Suppose a research group has an annotating budget <em>S </em>and a series of <em>K</em> questions they want answered. On the extremes, they can approach this in two ways:</p><ol><li><p>They can spend <em>S/K </em>dollars annotating data for each question they are interested in, and use DSL or an equivalent method to generalize annotations to the whole sample.</p></li><li><p>They can order the top K questions by interest and, for the top ones, spend the budget S annotating a sample large enough to conduct the analysis.</p></li></ol><p>Of course, an optimal strategy is probably something in between. But under which broad regime can science progress more efficiently? If the calibration methods are good enough, probably something closer to the first extreme. And there&#8217;s nothing uniquely &#8220;LM-ish&#8221; about this pipeline&#8212;the same logic would have applied in the BERT era, though the subset we needed to hand-label might have been larger then.</p><p>Second, we should clarify the threat model of malicious actors who may engage in bad science.</p><p>It&#8217;s worth remembering that we&#8217;ve been here before. The current anxiety around LM use echoes the <em>credibility revolution</em> in psychology and social science during the 2010s. Back then, the problem wasn&#8217;t that everyone was faking data (though a few were) but that researchers had too many degrees of freedom: selective reporting, flexible stopping rules, and post-hoc hypothesizing made false positives almost inevitable.</p><p>However, when it comes to curbing truly malicious use of LMs, I wonder whether LMs create &#8220;new degrees&#8221; of freedom for cheating that matter. After all, exclusions and minor tweaks to data that can impact the outcome were <em>always</em> possible. Maybe the threat of &#8220;<em>intentional&#8221;</em> LM hacking is that LMs create a scenario where this is truly impossible to track (without access to compute logs), but I wonder whether the delta here is substantial to matter (as, frankly, I guess most people could change their data and be fine, <a href="https://datacolada.org/109">as long as you don&#8217;t do it in Excel</a>).</p><p>Perhaps the threat is of a second-order nature, e.g., threatening the way we conduct science, making us <a href="https://www.nature.com/articles/s41586-025-09505-x">more lazy and dishonest when delegating tasks to LMs</a>, or ushering in a phase of scientific inquiry in which &#8220;<a href="https://www.nature.com/articles/s41586-024-07146-0">we produce more and understand less.</a>&#8221; But still, it seems to me that research addressing these threat models would provide the strongest evidence by finding evidence of said misuse, rather than hypothesizing it.</p><p>Third, I think we need to embrace the fact that human labels can also suck. And that failing to reproduce studies does not necessarily tell you that LMs are &#8220;wrong.&#8221;</p><p>In <a href="https://x.com/ey_985/status/1980272634458951998">Yang et al.</a>, it is worth noting that most of the annotated labels studied are a mix of human and previous ML methods (e.g., BERT). They find that interannotator agreement between strong LMs is around 0.6. But we&#8217;re not even sure what human interannotator agreement is across the tasks presented, to begin with. Authors find another report suggesting that n=20 studies reporting it have an average agreement of 0.73, but the figure smells of selection bias.</p><p>This ambiguity points to a deeper problem: when LM and human annotations diverge, we have no reliable way to know which side is wrong&#8212;or whether &#8220;wrong&#8221; even applies. Disagreement doesn&#8217;t reveal error so much as the instability of the categories themselves. In many social science tasks, ground truth is not discovered but negotiated through convention, as human annotation is too costly to do over and over again (just to measure variability). LMs are cheap, however, and may thus expose, rather than cause, the fuzziness of our constructs. Systematic gaps between human and model judgments can reflect algorithmic bias, but just as easily fatigue, inconsistency, or ideological drift among human coders. The uncomfortable reality is that for many annotation tasks, truth has always been approximate. Replication failures with LMs don&#8217;t necessarily prove the models are broken; they remind us that our human baselines were never as solid as we liked to think.</p><div><hr></div><p>My very centrist take is that labeling with LMs is neither trick nor treat. The real challenge is not replacement but calibration&#8212;whether we can model and report uncertainty with the same rigor we expect from any other instrument. LMs haven&#8217;t made annotation noisier; they&#8217;ve made its noise impossible to ignore. By exposing the randomness, inconsistency, and subjectivity that have long been hidden in human labeling, they force social science to confront its own measurement fragility. If we take that challenge seriously, LM labeling could make the field not lazier, but more transparent and scientific about its own uncertainty.</p><div><hr></div><p><strong>Acknowledgements</strong>: I thank Kristina Gligori&#263;, Paul R&#246;ttger, and Joachim Baumann for feedback on this blog post.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Missing Discipline in Computer Science]]></title><description><![CDATA[On Teaching Computer Scientists to Ask &#8220;Why&#8221;]]></description><link>https://doomscrollingbabel.manoel.xyz/p/the-missing-discipline-in-computer</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/the-missing-discipline-in-computer</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Sat, 04 Oct 2025 13:33:14 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/63053c1a-1a56-41c4-b93f-a880f69ad994_2304x1792.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Henderson&#8217;s first law of econometrics reads:</p><blockquote><p><em>When you read an econometric study done after 2005, the probability that the researcher has failed to take into account an objection that a non-economist will think of is close to zero. (<a href="https://www.econlib.org/hendersons-first-law-of-econometrics/">Source</a>)</em></p></blockquote><p>I dare to propose a similar law for machine learning:</p><blockquote><p><em>When an economist reads (and understands) an empirical </em>machine learning<em> study done after 2022, the probability that they will think of an objection that the researcher has failed to take into account is close to one.</em></p></blockquote><p>And why is that? Because the two fields treat empiricism in opposite ways. Econometrics was forged in the crucible of skepticism. Every paper is a defensive war against omitted variables, selection bias, and endogeneity. Its practitioners have been trained to see identification not as a formality, but as a matter of survival. Sit in on one of their seminars, and you&#8217;ll witness the kind of interrogation that, in Computer Science, would pass for an ambush.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>In machine learning, by contrast, the prevailing norm is demonstration, not falsification. Success is typically measured by predictive performance on a benchmark, rather than by causal clarity or robustness to alternative specifications. Yet, papers in machine learning increasingly treat models as behavioral systems to be studied experimentally (<a href="https://scholar.google.com/scholar?hl=en&amp;as_sdt=0%2C31&amp;q=%22Can+Large+Language+Model+Agents%22">&#8220;Can Large Language Models&#8230;&#8221;</a>). These papers make causal claims about systems&#8212;about how model behavior changes with prompts, environments, or training regimes&#8212;yet they timidly engage with the identification concerns that would accompany such claims in the empirical sciences.</p><p>This is not a machine learning problem alone&#8212;it permeates Computer Science, from HCI to Security and Privacy. Perhaps the reason lies in the field&#8217;s origins: a discipline born from the union of system building and theorem proving, where progress is demonstrated through construction, not systematic observation.</p><div><hr></div><p>So what? I don&#8217;t envy economists who spend five years publishing a paper. But I do believe Computer Science would benefit from adopting more of the empirical culture found in disciplines like economics and political science. The easiest place to start is education. Next semester, I&#8217;ll be teaching a graduate seminar titled <em>Empirical Research Methods for Computer Science</em>, partly because I want my students to have these tools&#8212;and partly because I want to ask a deeper question: what should the core empirics curriculum of Computer Science look like?</p><p>Below is my sketch, which I've spent a week working on.</p><p>First, we need to provide students with a vocabulary for causality. Many of the questions that computer scientists now confront&#8212;Does fine-tuning a model change its fairness? What is the effect of personalization on engagement?&#8212;are fundamentally causal questions. Yet few students have the conceptual tools to pose these questions precisely, let alone to answer them. Our students need a version of this material focusing on how to design credible experiments, reason about identification, and interpret evidence with humility.</p><p>Second, we need to teach students about regression. But not as a <em>tool for prediction, as</em> it is taught in machine learning courses. We need to teach them regression as a <em>tool for inference</em>: a way to estimate relationships, test hypotheses, and reason about uncertainty. Computer science students rarely learn to read a regression table, let alone to interrogate one. Yet these habits&#8212;thinking in terms of confounders, robustness checks, and sensitivity analyses&#8212;are what make empirical results interpretable and cumulative.</p><p>Third, we need to teach students about benchmarks. Benchmarks have long served as the de facto empirical infrastructure of Computer Science. Yet few students are ever taught to think critically about them. What does a benchmark actually measure?  When does improving performance reflect genuine scientific insight, and when does it reflect overfitting to a proxy task? Treating benchmarking as a scientific problem in its own right&#8212;concerned with validity, reliability, and construct design&#8212;would help students see that measurement is a form of theorizing about the world. </p><p>Fourth, we need to teach students about experimental design in the age of Computer Science. Our field now runs experiments at a scale and speed unimaginable in most other sciences, yet the principles remain the same: randomization, power, validity, and ethics. Students should learn how to design credible experiments, estimate statistical power, and ensure data quality in online settings&#8212;from platforms like Prolific to in-product A/B tests. These practices are not bureaucratic hurdles; they are what make empirical claims reproducible and trustworthy. </p><div><hr></div><p>None of these things is &#8220;hard,&#8221; and good material exists for a lot of them. Brady Neal has a great <a href="https://www.bradyneal.com/causal-inference-course">course</a> on causal inference; Moritz Hardt has an upcoming <a href="https://mlbenchmarks.org/">book</a> on benchmarks; and <a href="https://mixtape.scunning.com/">some</a> <a href="https://theeffectbook.net/">recent</a> Economics textbooks provide a very gentle introduction to causality and quasi-experimental methods. Yet for empirical rigor to take root in Computer Science, it needs to become part of how we train students to think&#8212;not just a set of optional skills. We need to signal that understanding identification, regression, and experimental design is as fundamental to being a computer scientist as knowing how to optimize an algorithm or prove a theorem. Only then can our field move from demonstration to explanation, and from performance to understanding.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Lean Chinese Room]]></title><description><![CDATA[Compression, Speed, and the Blurred Line of Comprehension]]></description><link>https://doomscrollingbabel.manoel.xyz/p/the-lean-chinese-room</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/the-lean-chinese-room</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Mon, 25 Aug 2025 14:54:52 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/742a9be3-bc45-4efb-9e15-8d60a9dfbea5_2304x1792.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The Chinese Room thought experiment is a go-to motivation for AI skeptics. It goes like this. Imagine a man locked in a room with a giant rulebook written in English. Slips of paper with Chinese characters are passed into the room. The man, who knows no Chinese, looks up the symbols in his rulebook, follows the instructions to the letter, and produces output characters that he passes back outside. To outsiders, the responses look fluent, even intelligent. But inside, the man is just shuffling symbols according to syntactic rules. There is no genuine understanding, no semantics, only blind symbol manipulation. (The original, by John Searle, is quite short; read it <a href="https://rintintin.colorado.edu/~vancecd/phil201/Searle.pdf">here</a>.)</p><p>Criticisms of this thought experiment are plentiful (I actually highly recommend the <a href="https://plato.stanford.edu/entries/chinese-room/#ReplChinRoomArgu">Stanford Encyclopedia of Philosophy</a>&#8217;s page on this). Here&#8217;s a quick summary: some argue that Searle mislocates understanding: while the man may not know Chinese, the system as a whole (man, books, and memory) might (the &#8220;Systems Reply&#8221;). Others suggest that running the program could generate a virtual mind distinct from the operator, which could be the true bearer of understanding (the &#8220;Virtual Mind Reply&#8221;). Embodiment theorists counter that meaning requires interaction with the world: put the program in a robot with sensors and effectors, and the symbols might gain genuine semantic grounding (the &#8220;Robot Reply&#8221;). Connectionists argue that if a program simulated the actual firing of neurons in a Chinese speaker&#8217;s brain, it would thereby instantiate understanding (the &#8220;Brain Simulator Reply&#8221;). Still others note that we attribute knowledge to other humans solely based on behavior; if a machine behaves equivalently, why not extend the same presumption (the &#8220;Other Minds Reply&#8221;)?</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>But what caught my eye are the arguments that question the Chinese Room argument due to its utter <em>inefficiency</em>. Namely, Steven Pinker in his 1997 book &#8220;How the Mind Works&#8221; writes:</p><blockquote><p><em>&#8216;The thought experiment slows down the waves to a range to which we humans no longer see them as light&#8230;. Similarly, Searle has slowed down the mental computations to a range in which we humans no longer think of it as understanding (since understanding is ordinarily much faster).&#8217;</em></p></blockquote><p>Similarly, in his 1980s paper "Fast Thinking" (I actually couldn&#8217;t find the original paper referenced in SEP), philosopher Daniel Dennett argues that the speed of processing is essential to any practical definition of intelligence:</p><blockquote><p><em>If you can&#8217;t figure out the relevant portions of the changing environment fast enough to fend for yourself, you are not practically intelligent, however complex you are&#8221;</em></p></blockquote><p>Another insightful paper by Daniel A. Wilkenfeld posits that  &#8220;<a href="https://link.springer.com/article/10.1007/s11098-018-1152-1">Understanding as compression</a>,&#8221; he writes:</p><blockquote><p><em>Understanding is a matter of compressing information about the understood so that it can be mentally useful. On this account, understanding amounts to having a representational kernel and the ability to use it to generate the information one needs regarding the target phenomenon.</em></p></blockquote><p>Reading this literature got me thinking of a &#8220;counter&#8221; thought experiment. Which I call &#8220;The Lean Chinese Room.&#8221; Imagine we depart from the exact room envisioned by Searle, but now, we start making things more complicated for the poor man who is trapped inside. Each day, we impose stricter conditions on him to generate his response.</p><ul><li><p>First, each day, the translation manuals are shortened. They still contain enough information for the man to generate the correct outputs, but in increasingly compressed form&#8212;abbreviated tables instead of exhaustive ones, general rules instead of endless lists of examples. The man is now forced to rely on compact kernels of information from which the rest can be reconstructed, rather than simply following brute-force, step-by-step recipes.</p></li><li><p>Second, there is now a timer in the room, which starts ticking the moment a translation request is slipped under the door. The man must respond before the bell rings. No longer can he spend hours leafing through massive binders to find the right match. He must become efficient, quick, and resourceful, producing his answers under the constraint of speed.</p></li></ul><p>My thought experiment is: <em>is there any moment at which the man must &#8220;understand&#8221; to be able to fulfill his duty?</em> At first, perhaps not. He can lean on the manuals and plod along, however slowly. But as the manuals shrink and the timer grows stricter, the burden shifts. To keep producing fluent answers, he must internalize shortcuts, general rules, and patterns that go beyond rote symbol-shuffling. At some point, the line between following instructions and genuine grasp blurs. If he can reconstruct meaning from compact cues and respond with speed and flexibility, we may be hard-pressed to deny that some form of understanding has emerged.</p><p>What, then, is he actually understanding? Perhaps not &#8220;Chinese&#8221; in the complete sense of an embodied speaker&#8212;he still has no taste of a hamburger or memory of Beijing. But he is no longer just pushing symbols blindly either. He has begun to understand the <em>system</em> itself: the patterns, abstractions, and compressed rules that allow Chinese to function as a communicative code. In other words, his performance reflects a structural or formal understanding, even if it is not the lived, grounded understanding Searle insists on. And suppose that structural understanding is enough to generate fluid, timely, context-sensitive responses. In that case, the boundary between mere manipulation and genuine comprehension becomes less clear than the original thought experiment suggests. The Lean Chinese Room thus complicates the stark form/meaning divide: under the pressures of compression and speed, &#8220;understanding&#8221; may emerge in degrees, not as an all-or-nothing property.</p><p>Okay, but maybe we strayed too far from the opening sentence, where I said &#8220;The Chinese Room thought experiment is a go-to motivation for AI skeptics.&#8221; Let me rewind. <a href="https://aclanthology.org/2020.acl-main.463/">Emily Bender and Alexander Koller (2020)</a> argue that large language models, trained purely on predicting sequences of words, cannot, in principle, acquire meaning. Their central claim is that meaning is the relation between linguistic form and communicative intent, grounded in the world. Models like BERT or GPT, in their view, only ever learn statistical regularities over form, no matter how impressive their outputs appear. They illustrate this with their now-famous &#8220;octopus test&#8221;: a clever octopus that eavesdrops on human conversations might learn to predict how speakers reply, but lacking any access to the world those utterances describe, it could not genuinely understand what the words mean&#65532;. In short, language models may mimic the surface of understanding, but because they are trained on form alone, they lack the crucial grounding that ties symbols to the world. In many ways, this is a restatement of the Chinese Room argument in contemporary form. Searle&#8217;s man shuffles symbols without knowing what they mean; Bender and Koller&#8217;s octopus predicts responses without ever grasping the world those responses refer to. Both thought experiments drive at the same intuition: fluency with form is not the same as grasp of meaning.</p><p>The Lean Chinese Room complicates this stark form/meaning divide. Like Bender and Koller&#8217;s octopus, the man in the room begins in a position of pure form manipulation. But as the manuals shrink and the timer ticks, his strategies must change: he is forced into compression, pattern extraction, and fast deployment. In other words, his behavior begins to look less like rote symbol-pushing and more like the internalization of structural knowledge. He still may not &#8220;know what a hamburger tastes like,&#8221; but he has acquired a functional, compressed kernel that lets him reconstruct meaning-like responses on demand. This suggests that understanding might not be a binary property requiring full world-grounding, but something that can emerge in degrees from the interplay of form, compression, and efficiency.</p><p>If that is right, then Bender and Koller&#8217;s sharp boundary between &#8220;form-only&#8221; systems and systems that achieve meaning may be too rigid. The Lean Chinese Room thought experiment shows how, under constraints of compression and speed, purely formal manipulation can drift into something we are tempted to call understanding, at least in the structural or functional sense. This does not refute the need for grounding altogether, but it weakens the claim that systems trained only on form are categorically barred from any understanding. Instead, it raises the possibility that there are <em>layers</em> of understanding: structural understanding rooted in form, and embodied understanding rooted in the world. Large language models may not have the latter, but they might well be on the path toward the former.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Retreating Human: God of the Gaps Logic in AI Resistance]]></title><description><![CDATA[I first came across the "God of the Gaps" concept when reading Richard Dawkins as an angsty teenager.]]></description><link>https://doomscrollingbabel.manoel.xyz/p/the-retreating-human-god-of-the-gaps</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/the-retreating-human-god-of-the-gaps</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Sun, 17 Aug 2025 12:15:19 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/13b7c5a0-9cd4-48bd-8058-50df25ebdd40_2304x1792.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I first came across the "God of the Gaps" concept when reading Richard Dawkins as an angsty teenager. Dawkins dedicates a whole chapter of <em>God Delusion </em>to complain about how something mysterious and previously seen as divine&#8212;lightning, disease, the complexity of the human eye&#8212;is finally explained by science. Some religious apologists will retreat to the next unexplained frontier as evidence of the existence of God. But this piece is not about religion, and criticisms of this tendency to attribute God to what we cannot yet explain have also bothered the pious. From his cell in WW2, German theologian Dietrich Bonhoeffer wrote:</p><blockquote><p><em>(...) How wrong it is to use God as a stopgap for the incompleteness of our knowledge. If, in fact, the frontiers of knowledge are being pushed further and further back (and that is bound to be the case), then God is being pushed back with them, and is therefore continually in retreat. We are to find God in what we know, not in what we don't know.</em></p></blockquote><p>This counter-argument, which was deeply embedded within the 2000s/2010s neo-atheism movement, couldn&#8217;t be more timely.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><strong>First, because AI is becoming &#8220;the hot&#8221; debate topic these days. </strong>There are debate subreddits dedicated to it (<a href="https://www.reddit.com/r/aiwars">r/aiwars</a>), visceral reactions abound, and intellectual careers are being built around the topic. But I can&#8217;t help but think that this is a debate far more nuanced and with far more consequences! People&#8217;s perspectives around AI are shaping geopolitics, infrastructure investments, and regulatory frameworks that will define the next decades. And unlike the question of god&#8217;s existence, which is fairly binary (at least from the perspective of an atheist), the question is now <em>whether</em> AI will influence society (it already does), but <em>how.</em></p><p><strong>Second, because bad arguments are rampant. </strong>Whether you are a <em>pro</em> or an <em>anti </em>(in r/aiwars linguo) is, in Gen Z speak, a <em>vibe</em>. And here we should note that hyping AI up can be very beneficial money-wise, and this was already the case far before GenAI (see Arvind and Sayash&#8217;s book for plenty of examples): a timeline where AI proves revolutionary is good news for the valuation of companies, from startups to behemoths. And in that context, every breakthrough gets breathlessly covered as "AGI is here!" and every limitation gets dismissed as "just a scaling problem.&#8221;</p><p>But this post is a criticism of a specific pattern of AI skepticism, which is the reverse: <strong>seeing humanness in the next frontier that is unconquered by AI. </strong>Before &#8220;delving&#8221; further, let me just make clear why I think criticizing &#8220;<em>anti-AI&#8221;</em> takes is essential. AI is genuinely consequential technology that's already reshaping labor markets, educational systems, and creative industries in measurable ways. This means we urgently need clear thinking about how society functions in an AI-saturated world. And we do not get to help shape the world for the better if debates center around what AI cannot yet do.</p><h3>Human of the Gaps and the Stochastic Parrot</h3><p>Back in 2021, Bender et al. published the very influential &#8220;<a href="https://dl.acm.org/doi/10.1145/3442188.3445922">On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? &#129436;</a>&#8221;. This paper provides essential lingo for what I&#8217;m calling the <em>&#8220;human of the gaps&#8221;</em> argumentative strategy. Authors define Language Models as generating of "haphazardly stitch together sequences of linguistic forms" based on probabilistic patterns from vast training data. But they do so "without any reference to meaning"; the text they generate is not grounded in "communicative intent, any model of the world, or any model of the reader's state of mind."</p><p>The paper arguments are not very convincing today. Current LMs can solve unseen Math Olympiad problems through complex &#8220;reasoning chains&#8221; (<a href="https://deepmind.google/discover/blog/advanced-version-of-gemini-with-deep-think-officially-achieves-gold-medal-standard-at-the-international-mathematical-olympiad/">see Google&#8217;s official IMO gold</a>), they definitely can create world models of the underlying data (<a href="https://arxiv.org/pdf/2210.13382">see the Othello paper</a>), and they can act as agents that autonomously use tools to accomplish goals (<a href="https://arxiv.org/abs/2210.03629">see ReAct</a>). Yet, &#8220;true&#8221; understanding, intelligence, and consciousness are so hard to define (and measure), that their achievement can be always moved to the next unresolved frontier.</p><p>This is where the &#8220;Human-of-the-Gaps&#8221; move emerges (or perhaps <a href="https://www.astralcodexten.com/p/my-bet-ai-size-solves-flubs">&#8220;The Marcus maneuver&#8221;</a>?). Every time a frontier falls, skeptics retreat to the next one, insisting that is where &#8220;real&#8221; humanness lies. Machines beat us at chess? Well, chess is just calculation. They learn to translate across dozens of languages? Translation is pattern-matching. They generate music, code, art, and even pass professional licensing exams? These, we&#8217;re told, are still not &#8220;true&#8221; creativity or &#8220;true&#8221; reasoning. The essence of being human always lives one step further into the fog, wherever technology has not yet penetrated.</p><p>The problem with this move is not that it is <em>wrong</em> to notice gaps. Gaps are real and important. The problem is that it becomes a kind of motte-and-bailey: it defines &#8220;human&#8221; precisely as &#8220;whatever machines cannot do (yet).&#8221; That definition is, by construction, unfalsifiable. It is also unhelpful for guiding real-world decisions about governance, labor, or ethics. If we only anchor our self-worth and policy frameworks in what remains <em>unachieved</em>, then every technical advance destabilizes the foundation of the argument.</p><p>And this matters because while we bicker over whether models &#8220;really&#8221; understand, their measurable consequences accumulate. Generative models are not waiting for philosophical consensus to reshape education, journalism, software engineering, and creative industries. They are already altering workflows, wages, and the distribution of expertise. To insist that these systems are &#8220;mere parrots&#8221; is to miss the obvious: parrots that can code, draft contracts, or prove theorems are already socially transformative, regardless of whether their internal processes pass your metaphysical bar for &#8220;understanding.&#8221;</p><h3>Another Path Forward</h3><p>What should we do instead? Let&#8217;s turn again to Bender et al. (2021) and not throw the baby out with the bathwater. They made important contributions about bias amplification, environmental costs, and deployment risks that remain highly relevant. Yet, these contributions were framed based on <strong>what AI was at the time of writing</strong>, not about what it was not.</p><p>This is the crucial difference. A productive critique does not define itself around what the technology currently lacks, but rather around the concrete harms, limitations, and externalities that can be measured and governed. We don&#8217;t need to speculate about &#8220;true&#8221; consciousness to see that large-scale models are shaping energy policies around the world, that training data often encodes social prejudices, or that downstream deployment can exacerbate inequality in labor markets. Those are tractable, empirical claims, and they remain urgent regardless of whether models someday cross into capacities we would consider &#8220;understanding.&#8221;</p><p>Another path forward, then, is to build an evaluative vocabulary around what systems demonstrably do, not what they metaphysically are. This aligns with Narayanan and Kapoor&#8217;s proposal to view <a href="https://knightcolumbia.org/content/ai-as-normal-technology">AI as &#8220;normal technology&#8221;</a> like electricity or the internet: transformative, yes, but not a <em>sui generis</em> alien species. In their words, &#8220;to view AI as normal is not to understate its impact&#8212;even transformative, general-purpose technologies such as electricity and the internet are &#8216;normal&#8217; in our conception. But it is in contrast to both utopian and dystopian visions (&#8230;) which treat it akin to a separate species&#8221; &#65532;. Framing AI in this way shifts the focus from ontological riddles (whether models &#8220;really&#8221; understand or &#8220;truly&#8221; reason) to practical governance. It means centering debates on capabilities, externalities, and failure modes: what tasks these systems can perform and how reliably, what costs they impose on labor markets and the environment, and how they fail under distributional shifts or adversarial pressure.</p><p>Anchoring debates this way helps us avoid the endless retreat into &#8220;Human-of-the-Gaps&#8221; territory, and instead grounds our choices in observable consequences. It also gives us a way to govern these technologies: to measure, benchmark, red-team, and legislate, rather than waiting for some metaphysical threshold to be crossed. Prediction and speculation are not useless&#8212;they help us scan the horizon&#8212;but they are no substitute for evidence. When speculation hardens into certainty, it distorts priorities: risks that are near and measurable get neglected, while imagined futures soak up attention.</p><p>Bonhoeffer&#8217;s advice scales: find our footing in what we know. For AI, that means resisting both the theological impulse to search for a sacred remainder machines can never touch and the eschatological impulse to declare transcendence at each leaderboard bump. The frontier will keep moving; that is the nature of science and technology. What must not is our sense of value. Instead of tying human worth to whichever frontier machines have not yet crossed, we should tie it to responsibility, justice, and the institutions we build.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[How AI May Change Social Media]]></title><description><![CDATA[Everyone is wrong about AI Slop #3]]></description><link>https://doomscrollingbabel.manoel.xyz/p/how-ai-may-change-social-media</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/how-ai-may-change-social-media</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Sun, 03 Aug 2025 20:55:40 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4780d778-6b63-444c-8ee6-9179948294f6_2304x1792.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The question of AI-generated content and social media is, if nothing else, an invitation to dust off your favorite economic metaphors and let them parade in an algorithmic carnival. Let&#8217;s call social media a market, a digital bazaar where the wares are content, paid not in dollars but in swipes, likes, and half-listened seconds. In this market, &#8220;producers&#8221; are the content creators, and the rest of us are &#8220;The Great Unwashed,&#8221; eternally scrolling.</p><p>Of course, this was not always so, back in the sepia-tinted 2010s, when Facebook statuses flourished and your cousin's photos jockeyed for screen space with existential memes and early attempts at going viral. Everyone was both producer and consumer, a kind of content hobbyist, if you will. Then, as if overnight, short-form video took online platforms by storm, and the producer-consumer distinction ossified.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>A significant factor here could be what <a href="https://www.ystrickler.com/the-dark-forest-theory-of-the-internet/">Strickler</a> refers to as &#8220;The Dark Forest Theory of the Web.&#8221; Turns out that being a content hobbyist on the Web had more downsides than upsides. As public feeds became money-making machines where shitposting had real-world consequences, many ordinary users retreated into quieter corners: closed groups, encrypted chats, Discord servers, obscure substacks.</p><p>Now, we find ourselves in the era of the Content Creator as businesses. Mr. Beast presides over a minor fiefdom of editors, thumbnail artists, and data-obsessed consigliere. But for these crowned heads of engagement, the real bottleneck isn&#8217;t a lack of ideas, money, or labor. It&#8217;s the invisible hand of platform design. Post 100 Mr. Beast videos in a single week and you&#8217;ll discover, painfully, that YouTube&#8217;s algorithm is less a free market and more a jealous bureaucrat, suspicious of overproduction and quick to impose rationing. The supply of content is limited not by the creator&#8217;s capacity, but by the platform&#8217;s appetite for spectacle.</p><p>What, then, of AI? Will a bevy of GPTs and DALL-Es let the top creators dominate even more completely, flooding the feed with a ceaseless slurry of Beastly wonders? Not really. AI may offer a few more pixels of polish, a higher ratio of jump cuts per minute, or some extra flair in the thumbnail. Still, this is only a slightly better mousetrap for those who have already captured the cheese. The platforms&#8217; monarchs might squeeze out a bit more efficiency or visual punch, but AI won&#8217;t fundamentally upend the pecking order at the top.</p><p>If there is a revolution, it won&#8217;t start in the castle, but among the rabble at the gates. Because here&#8217;s where things get interesting: the so-called &#8220;long tail.&#8221; Some forms of content, think daily vlogs or dog photos, were always democratized. Still, others, like animation or high-production visual storytelling, remained the domain of the technically gifted, the well-funded, or the terminally stubborn.</p><p>AI promises to decrease the cost of content creation and the skills required. What once took a studio can now be managed by a hobbyist and a well-tuned prompt. Text-to-image generators, auto-editors, voice clones, and AI-powered storytellers turn what was once artisanal labor into the push of a button. Animation for the masses. Deepfakes for the family group chat. Explainers in any language for any niche, no matter how obscure or underloved.</p><p>We may soon find ourselves awash in creative micro-genres and hobbyist enclaves that would never have survived in the old regime: AI-generated history documentaries narrated by your favorite Twitch streamer, animated music videos starring your D&amp;D character, explainer series about obscure philosophy, all built from scratch in a bedroom. AI won&#8217;t just automate what already exists; it will open doors to what was not profitable to create.</p><p>But as the means of creation become almost frictionless, a new paradox emerges: when anyone can make anything, what matters isn&#8217;t just the thing itself. Brand, personality, community, these are already moats. People don&#8217;t crave an endless buffet of faceless content; they want context, connection, a human story behind the pixels. Parasocial relationships, the one-sided love affairs between creator and audience, are the secret ingredient in the attention economy. It may be that the more abundant the content becomes, the scarcer authenticity and presence will feel.</p><p>So don&#8217;t expect the web to be swamped by faceless AI content. The value of being a person may go up as the price of everything else approaches zero.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Will Not Eat Social Media]]></title><description><![CDATA[Everyone is wrong about AI Slop #2]]></description><link>https://doomscrollingbabel.manoel.xyz/p/ai-will-not-eat-social-media</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/ai-will-not-eat-social-media</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Fri, 04 Jul 2025 19:43:54 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/78eece0c-4715-4a9d-9203-9054419d1662_2304x1792.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Generative AI is poised to reshape social media dramatically&#8212;no question about it (is anyone else super mindful of using &#8220;&#8212;&#8221; these days?). From content creation to recommendation algorithms, its influence will be far-reaching. But contrary to the more <a href="https://www.trend-mill.com/p/ai-sludge-will-rot-social-media">alarmist</a> takes, I believe that AI will not &#8220;destroy&#8221; or &#8220;eat&#8221; social media as we know it. Especially the kind of AI-generated content that exists today. This prediction overlooks what we know about people&#8217;s social media experiences and the creator economy. Namely, they overlook three things, which I discuss in the essay.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Authenticity is the currency of social media creators</h1><p>The idea that AI will overtake social media assumes that users will be satisfied consuming content generated entirely by machines. This is highly unlikely, considering how influencer-driven social media platforms like TikTok and YouTube are currently. Influencers thrive because users develop parasocial relationships with their creators, they follow their lives, care about their opinions, and participate in their narratives. The keyword here is &#8220;authenticity,&#8221; which AI slop is anything but.</p><p>But how about AI influencers? Yes, one could imagine a world where people form connections with fictitious characters, similar to how Japan&#8217;s virtual idols like <a href="https://en.wikipedia.org/wiki/Hatsune_Miku">Hatsune Miku</a> have gained devoted fanbases. But this kind of content is carefully crafted. Miku isn&#8217;t just a product of automation; she&#8217;s the result of intentional worldbuilding, aesthetic curation, and human creativity.  </p><p>The &#8220;AI will eat social media&#8221; discourse implies that sheer scale will tip the balance. We can now generate endless amusing content! But scale is at odds with authenticity. Even in the absence of the very real anti-AI sentiment that exists online, the fact that content is produced&nbsp;<em>en masse</em>, without exceeding care or thought, will make it very hard for people to create the emotional connections that drive the online spaces.</p><h1>Social media users are not passive consumers of content</h1><p>Social media is not a one-size-fits-all broadcast model. It&#8217;s a sprawling, fragmented ecosystem made up of micro-niches, subcultures, fandoms, and highly individualized streams of content. Unlike traditional media, which pushes the same content to everyone, social media is shaped by the cumulative actions and preferences of each user. People don&#8217;t just consume content. They <em>curate</em> their own feeds through what they click on, follow, like, ignore, comment on, and share. In doing so, they effectively place themselves into specific corners of the internet, each with its own norms, aesthetics, values, and inside jokes.</p><p>These corners aren&#8217;t just algorithmically generated&#8212;they&#8217;re socially constructed. BookTok isn&#8217;t just &#8220;videos about books,&#8221; and Skincare YouTube isn&#8217;t just &#8220;product reviews.&#8221; These niches become communities, defined by shared language, common reference points, and a sense of belonging. Participation requires a kind of cultural literacy. You have to know how to speak the dialect, what trends are current, what behaviors are rewarded, and what signals authenticity.</p><p>Users build their content diets through a <em>potpourri </em>of &#8220;niches,&#8221; and it is just implausible that GenAI will eat all of them. In some &#8220;types&#8221; of content,  I&#8217;d argue that using GenAI may be more of a liability than a disadvantage. Sure, AI will likely be used as a tool to create memes, but will it be helpful for content rooted in lived experiences?</p><h1><strong>Social media is built to handle enormous amounts of content</strong></h1><p>The &#8220;AI will eat social media&#8221; narrative treats scale as something novel or destabilizing, when in reality, social media platforms have been built to handle overwhelming volumes of content. TikTok, YouTube, Instagram, X, and Reddit are not overwhelmed by too much content; they are <em>defined</em> by it. These systems are designed from the ground up to process, sort, and surface content in real time. Most of what gets uploaded&#8212;whether human-generated or AI-generated&#8212;is of low quality, ignored, and quickly buried. That&#8217;s business as usual.</p><p>The fact that people can now use AI to generate large volumes of content doesn&#8217;t fundamentally change how these platforms operate. If users don&#8217;t engage with it, it simply won&#8217;t circulate. And if they <em>do</em> engage with it, then sure, platforms will show them more of it. But they won&#8217;t show <em>only</em> that kind of content, because as discussed earlier, users aren&#8217;t passive consumers. They seek out relevance, novelty, personality, and cultural context. A wall of AI-generated &#8220;dank memes&#8221; might go viral, but it won&#8217;t define everyone&#8217;s social media experience. </p><div><hr></div><p>GenAI will almost certainly reshape the texture of social media, but it won&#8217;t erase the dynamics that make these platforms compelling to begin with. Social media is shaped by human behavior, taste, identity, and community. Authenticity still matters. User agency still matters. And the infrastructure is already built to filter the flood. AI might change the <em>how</em> of content creation, but it won&#8217;t change the <em>why</em> of engagement. For better or worse, social media is still powered by people&#8212;and that&#8217;s not going away anytime soon.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI-Generated Videos are Actually Quite Expensive to Create]]></title><description><![CDATA[Everyone is wrong about AI Slop #1]]></description><link>https://doomscrollingbabel.manoel.xyz/p/ai-generated-videos-are-actually</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/ai-generated-videos-are-actually</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Tue, 01 Jul 2025 19:50:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/b9f15466-a382-4291-a521-ee3a5d0d7da8_2304x1792.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>Everyone is wrong about AI Slop.</strong></p><p>Right there, I said it. AI-generated content has become a mainstream topic (as evidenced by this <a href="https://www.youtube.com/watch?v=TWpg1RmzAbc&amp;sttick=0">John Oliver</a> video about it), but the public discourse surrounding it is a bit detached from reality. In the next few months, I plan to write a series of posts that comment and add nuance to some of the underlying claims informing public discourse around AI-generated content. As I do so, I hope to shed light on the whole AI content creation ecosystem, which I have &#8220;delved deep into&#8221; in the last few weeks. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Today, let me take a stab at:</p><blockquote><p><em>&#8220;AI-generated content is cheap to create.&#8221;</em></p></blockquote><p>This is demonstrably false for video, which currently costs a significant amount (as of June 2025). For example, consider the very amusing cat olympics video:</p><div id="tiktok-iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40pablo.prompt%2Fvideo%2F7517691413849492758&amp;key=e27c740634285c9ddc20db64f73358dd" class="tiktok-wrap outer" data-attrs="{&quot;url&quot;:&quot;https://www.tiktok.com/@pablo.prompt/video/7517691413849492758&quot;,&quot;title&quot;:&quot;Michi Olimpiadas &#128008;&#128008;&#8205;&#11035; V&#237;deos creados con un nuevo modelo de IA que, como pueden ver, logra unas f&#237;sicas incre&#237;bles: Hailuo 02 de @Hailuo AI (MiniMax)  Denle amor que si les gusta traer&#233; m&#225;s videos creados con Hailuo de las olimpiadas gatunas &#128524;&#128571; #hailuo #minimax #veo3gratis #veo3 #gatos #michi #olimpiadas &quot;,&quot;thumbnail_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e2d1aea0-6a1a-48c3-9cdb-05ad2f3b7019_1080x1920.jpeg&quot;,&quot;author&quot;:&quot;Pablo Prompt&quot;,&quot;embed_url&quot;:&quot;https://cdn.iframe.ly/api/iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40pablo.prompt%2Fvideo%2F7517691413849492758&amp;key=e27c740634285c9ddc20db64f73358dd&quot;,&quot;author_url&quot;:&quot;https://www.tiktok.com/@pablo.prompt&quot;,&quot;belowTheFold&quot;:false}" data-component-name="TikTokCreateTikTokEmbed"><iframe id="iframe-tiktok-iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40pablo.prompt%2Fvideo%2F7517691413849492758&amp;key=e27c740634285c9ddc20db64f73358dd" class="tiktok-iframe" src="https://cdn.iframe.ly/api/iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40pablo.prompt%2Fvideo%2F7517691413849492758&amp;key=e27c740634285c9ddc20db64f73358dd" frameborder="0" allow="autoplay; fullscreen; encrypted-media" allowfullscreen="" scrolling="no"></iframe><iframe src="https://team-hosted-public.s3.amazonaws.com/set-then-check-cookie.html" id="third-party-iframe-tiktok-iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40pablo.prompt%2Fvideo%2F7517691413849492758&amp;key=e27c740634285c9ddc20db64f73358dd" class="third-party-cookie-check-iframe" style="display: none;"></iframe><div class="tiktok-wrap static" data-component-name="TikTokCreateStaticTikTokEmbed"><a href="https://www.tiktok.com/@pablo.prompt/video/7517691413849492758" target="_blank"><img class="tiktok thumbnail" src="https://substackcdn.com/image/fetch/$s_!t4XH!,w_640,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d1aea0-6a1a-48c3-9cdb-05ad2f3b7019_1080x1920.jpeg" style="background-image: url(https://substackcdn.com/image/fetch/$s_!t4XH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d1aea0-6a1a-48c3-9cdb-05ad2f3b7019_1080x1920.jpeg);"></a><div class="content"><a class="author" href="https://www.tiktok.com/@pablo.prompt" target="_blank">@pablo.prompt</a><a class="title" href="https://www.tiktok.com/@pablo.prompt/video/7517691413849492758" target="_blank">Michi Olimpiadas &#128008;&#128008;&#8205;&#11035; V&#237;deos creados con un nuevo modelo de IA que, como pueden ver, logra unas f&#237;sicas incre&#237;bles: Hailuo 02 de @Hailuo AI (MiniMax)  Denle amor que si les gusta traer&#233; m&#225;s videos creados con Hailuo de las olimpiadas gatunas &#128524;&#128571; #hailuo #minimax #veo3gratis #veo3 #gatos #michi #olimpiadas </a></div></div><div class="fallback-failure" id="fallback-failure-tiktok-iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40pablo.prompt%2Fvideo%2F7517691413849492758&amp;key=e27c740634285c9ddc20db64f73358dd"><div class="error-content"><img class="error-icon" src="https://substackcdn.com//img/alert-circle.svg">Tiktok failed to load.<br><br>Enable 3rd party cookies or use another browser</div></div></div><p>This has 34 seconds. As of the time of writing, access to the model (likely) used here (Veo 3) can be obtained either through a monthly subscription costing $250 or via the official API. The subscription provides 12,500 credits, and generating a single 8-second video costs 150 credits. This means that a video sequence costs three bucks. (Using the API costs twice as much as 0.75 cents per second of video).</p><p>The video above has six different scenes, each probably created independently. Meaning it costs $18 to make it. But hey, it is implausible that everything worked perfectly the first time. Let us conservatively assume that for each take included in the final cut, the creator made four videos, meaning that three videos did not make the final version. This means this video costs $72.</p><p>Let us assume a relatively high &#8220;Revenue Per Mille,&#8221; i.e., how much money TikTok pays you for 1,000 views. Specifically, let's assume someone can earn 70 cents per 1,000 views. To offset costs, they need their videos to reach a little over 100,000 people. Which, frankly, is relatively high. I don&#8217;t have the exact figures, but I&#8217;m pretty sure that the vast majority of videos achieve&nbsp;<em>fewer</em> than 100,000 views. I&#8217;d be surprised if the median person using Veo3 isn&#8217;t losing money.</p><h2>Running on VC Money</h2><p>This begs the question. If video is so expensive, why are we seeing so much AI-generated video everywhere? And I&#8217;d argue that what is going on here is that:</p><ol><li><p>A bunch of people are indeed losing money doing AI-generated content.</p></li><li><p>Few people are making a significant amount of money from it.</p></li><li><p>The cost of producing AI-generated content is reduced by the fact that numerous startups are burning money to enable people to create content.</p></li></ol><p>Remember when Uber just started, and it was incredibly cheap to Uber anywhere? I believe this is what is happening here. I&#8217;d argue that few of the AI content creation people actually use the top tools offered by Google or OpenAI. Indeed, if you click on one of the thousands of how-to generate AI tutorial videos, you will probably hear of a bunch of different AI generation platforms: Vadoo, Mirage, Leonardo AI. </p><p>Most of these platforms do not have an official API and operate on a &#8220;You get <em>k</em> credits a month&#8221; basis. These are probably AI startups operating at a loss. E.g., Leonardo AI offers Google&#8217;s Veo3 <a href="https://leonardo.ai/veo-3/">cheaper than Google</a><em><strong>:</strong></em></p><blockquote><p>Leonardo.Ai offers a lower cost of entry to access Veo 3, from just $10 USD* per month. Video generation with Veo 3 is more affordable on the Leonardo platform, with a lower equivalent dollar cost of total tokens per generation ($0.75 on Google&#8217;s platform vs. approximately $0.30 on Leonardo.Ai).</p></blockquote><h2>A Market Perspective</h2><p>Okay, but let&#8217;s take a step back. What does it mean for something to be cheap? The cat Olympics video has 3 million views, meaning it might have earned someone $2,000 (assuming it is monetized). And frankly, <em>video-generated content will only become cheaper</em>, so it will probably cost only $0.70 to make one of those in a year or two. So why am I still saying &#8220;AI-generated content is not cheap&#8221;? Because creating (some types of) content on TikTok is already pretty cheap.</p><div id="tiktok-iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40khaby.lame%2Fvideo%2F6978110267225935110%3Fembed_source%3D121374463%252C121468991%252C121439635%252C121749182%252C121433650%252C121404359%252C121497414%252C121477481%252C121351166%252C121772354%252C121487028%252C121679410%252C121331973%252C120811592%252C120810756%252C121503376%253Bnull%253Bembed_masking%26refer%3Dembed%26referer_url%3Dcdn.iframe.ly%252Fapi%252Fiframe%253Fmedia%253D1%2526app%253D1%2526url%253Dhttps%25253A%25252F%25252Fwww.tiktok.com%25252F%252540khaby.lame%25252Fvideo%25252F6972535556127444229%25253Flang%25253Den%2526key%253De27c740634285c9ddc20db64f73358dd%26referer_video_id%3D6972535556127444229&amp;key=e27c740634285c9ddc20db64f73358dd" class="tiktok-wrap outer" data-attrs="{&quot;url&quot;:&quot;https://www.tiktok.com/@khaby.lame/video/6978110267225935110&quot;,&quot;title&quot;:&quot;Nooo guys Please nooooo &#128557;&#128557;&#128557; &#129315;#learnfromkhaby #learnwithtiktok #ImparaConTikTok&quot;,&quot;thumbnail_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bb333dd7-d3f0-41eb-af04-a37e09b132da_720x1280.jpeg&quot;,&quot;author&quot;:&quot;Khabane lame&quot;,&quot;embed_url&quot;:&quot;https://cdn.iframe.ly/api/iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40khaby.lame%2Fvideo%2F6978110267225935110%3Fembed_source%3D121374463%252C121468991%252C121439635%252C121749182%252C121433650%252C121404359%252C121497414%252C121477481%252C121351166%252C121772354%252C121487028%252C121679410%252C121331973%252C120811592%252C120810756%252C121503376%253Bnull%253Bembed_masking%26refer%3Dembed%26referer_url%3Dcdn.iframe.ly%252Fapi%252Fiframe%253Fmedia%253D1%2526app%253D1%2526url%253Dhttps%25253A%25252F%25252Fwww.tiktok.com%25252F%252540khaby.lame%25252Fvideo%25252F6972535556127444229%25253Flang%25253Den%2526key%253De27c740634285c9ddc20db64f73358dd%26referer_video_id%3D6972535556127444229&amp;key=e27c740634285c9ddc20db64f73358dd&quot;,&quot;author_url&quot;:&quot;https://www.tiktok.com/@khaby.lame&quot;,&quot;belowTheFold&quot;:true}" data-component-name="TikTokCreateTikTokEmbed"><iframe id="iframe-tiktok-iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40khaby.lame%2Fvideo%2F6978110267225935110%3Fembed_source%3D121374463%252C121468991%252C121439635%252C121749182%252C121433650%252C121404359%252C121497414%252C121477481%252C121351166%252C121772354%252C121487028%252C121679410%252C121331973%252C120811592%252C120810756%252C121503376%253Bnull%253Bembed_masking%26refer%3Dembed%26referer_url%3Dcdn.iframe.ly%252Fapi%252Fiframe%253Fmedia%253D1%2526app%253D1%2526url%253Dhttps%25253A%25252F%25252Fwww.tiktok.com%25252F%252540khaby.lame%25252Fvideo%25252F6972535556127444229%25253Flang%25253Den%2526key%253De27c740634285c9ddc20db64f73358dd%26referer_video_id%3D6972535556127444229&amp;key=e27c740634285c9ddc20db64f73358dd" class="tiktok-iframe" src="https://cdn.iframe.ly/api/iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40khaby.lame%2Fvideo%2F6978110267225935110%3Fembed_source%3D121374463%252C121468991%252C121439635%252C121749182%252C121433650%252C121404359%252C121497414%252C121477481%252C121351166%252C121772354%252C121487028%252C121679410%252C121331973%252C120811592%252C120810756%252C121503376%253Bnull%253Bembed_masking%26refer%3Dembed%26referer_url%3Dcdn.iframe.ly%252Fapi%252Fiframe%253Fmedia%253D1%2526app%253D1%2526url%253Dhttps%25253A%25252F%25252Fwww.tiktok.com%25252F%252540khaby.lame%25252Fvideo%25252F6972535556127444229%25253Flang%25253Den%2526key%253De27c740634285c9ddc20db64f73358dd%26referer_video_id%3D6972535556127444229&amp;key=e27c740634285c9ddc20db64f73358dd" frameborder="0" allow="autoplay; fullscreen; encrypted-media" allowfullscreen="" scrolling="no" loading="lazy"></iframe><iframe src="https://team-hosted-public.s3.amazonaws.com/set-then-check-cookie.html" id="third-party-iframe-tiktok-iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40khaby.lame%2Fvideo%2F6978110267225935110%3Fembed_source%3D121374463%252C121468991%252C121439635%252C121749182%252C121433650%252C121404359%252C121497414%252C121477481%252C121351166%252C121772354%252C121487028%252C121679410%252C121331973%252C120811592%252C120810756%252C121503376%253Bnull%253Bembed_masking%26refer%3Dembed%26referer_url%3Dcdn.iframe.ly%252Fapi%252Fiframe%253Fmedia%253D1%2526app%253D1%2526url%253Dhttps%25253A%25252F%25252Fwww.tiktok.com%25252F%252540khaby.lame%25252Fvideo%25252F6972535556127444229%25253Flang%25253Den%2526key%253De27c740634285c9ddc20db64f73358dd%26referer_video_id%3D6972535556127444229&amp;key=e27c740634285c9ddc20db64f73358dd" class="third-party-cookie-check-iframe" style="display: none;" loading="lazy"></iframe><div class="tiktok-wrap static" data-component-name="TikTokCreateStaticTikTokEmbed"><a href="https://www.tiktok.com/@khaby.lame/video/6978110267225935110" target="_blank"><img class="tiktok thumbnail" src="https://substackcdn.com/image/fetch/$s_!g-I9!,w_640,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb333dd7-d3f0-41eb-af04-a37e09b132da_720x1280.jpeg" style="background-image: url(https://substackcdn.com/image/fetch/$s_!g-I9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbb333dd7-d3f0-41eb-af04-a37e09b132da_720x1280.jpeg);" loading="lazy"></a><div class="content"><a class="author" href="https://www.tiktok.com/@khaby.lame" target="_blank">@khaby.lame</a><a class="title" href="https://www.tiktok.com/@khaby.lame/video/6978110267225935110" target="_blank">Nooo guys Please nooooo &#128557;&#128557;&#128557; &#129315;#learnfromkhaby #learnwithtiktok #ImparaConTikTok</a></div></div><div class="fallback-failure" id="fallback-failure-tiktok-iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40khaby.lame%2Fvideo%2F6978110267225935110%3Fembed_source%3D121374463%252C121468991%252C121439635%252C121749182%252C121433650%252C121404359%252C121497414%252C121477481%252C121351166%252C121772354%252C121487028%252C121679410%252C121331973%252C120811592%252C120810756%252C121503376%253Bnull%253Bembed_masking%26refer%3Dembed%26referer_url%3Dcdn.iframe.ly%252Fapi%252Fiframe%253Fmedia%253D1%2526app%253D1%2526url%253Dhttps%25253A%25252F%25252Fwww.tiktok.com%25252F%252540khaby.lame%25252Fvideo%25252F6972535556127444229%25253Flang%25253Den%2526key%253De27c740634285c9ddc20db64f73358dd%26referer_video_id%3D6972535556127444229&amp;key=e27c740634285c9ddc20db64f73358dd"><div class="error-content"><img class="error-icon" src="https://substackcdn.com//img/alert-circle.svg" loading="lazy">Tiktok failed to load.<br><br>Enable 3rd party cookies or use another browser</div></div></div><p>Let&#8217;s contrast AI-generated video with another dominant format on TikTok: reacts. Videos with stitched reactions can be made in minutes on a phone, for free. These require zero API tokens, no model prompts, and no rendering time. And if you&#8217;re recycling existing media (which many viral TikToks do), your marginal production cost is effectively zero. Yet, you can still get &#8220;Internet famous&#8221; from doing reacts&#8212;get sponsorships, brand deals, make merch, etc. Khaby Lame, the world&#8217;s most popular TikTok, kinda did it.</p><div id="tiktok-iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40fooooxmovie%2Fvideo%2F7278236490289794334%3Fq%3Dmovies%26t%3D1751397910785&amp;key=e27c740634285c9ddc20db64f73358dd" class="tiktok-wrap outer" data-attrs="{&quot;url&quot;:&quot;https://www.tiktok.com/@fooooxmovie/video/7278236490289794334&quot;,&quot;title&quot;:&quot;Part 1 #thewolfofwallstreet #leonardodicaprio #jonahhill #margotrobbie #matthewmcconaughey #movie #film #fyp #foryou &quot;,&quot;thumbnail_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2965e1e1-69bf-40b6-a334-55e227eb0d57_1080x1920.jpeg&quot;,&quot;author&quot;:&quot;Fooooxmovie&quot;,&quot;embed_url&quot;:&quot;https://cdn.iframe.ly/api/iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40fooooxmovie%2Fvideo%2F7278236490289794334%3Fq%3Dmovies%26t%3D1751397910785&amp;key=e27c740634285c9ddc20db64f73358dd&quot;,&quot;author_url&quot;:&quot;https://www.tiktok.com/@fooooxmovie&quot;,&quot;belowTheFold&quot;:true}" data-component-name="TikTokCreateTikTokEmbed"><iframe id="iframe-tiktok-iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40fooooxmovie%2Fvideo%2F7278236490289794334%3Fq%3Dmovies%26t%3D1751397910785&amp;key=e27c740634285c9ddc20db64f73358dd" class="tiktok-iframe" src="https://cdn.iframe.ly/api/iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40fooooxmovie%2Fvideo%2F7278236490289794334%3Fq%3Dmovies%26t%3D1751397910785&amp;key=e27c740634285c9ddc20db64f73358dd" frameborder="0" allow="autoplay; fullscreen; encrypted-media" allowfullscreen="" scrolling="no" loading="lazy"></iframe><iframe src="https://team-hosted-public.s3.amazonaws.com/set-then-check-cookie.html" id="third-party-iframe-tiktok-iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40fooooxmovie%2Fvideo%2F7278236490289794334%3Fq%3Dmovies%26t%3D1751397910785&amp;key=e27c740634285c9ddc20db64f73358dd" class="third-party-cookie-check-iframe" style="display: none;" loading="lazy"></iframe><div class="tiktok-wrap static" data-component-name="TikTokCreateStaticTikTokEmbed"><a href="https://www.tiktok.com/@fooooxmovie/video/7278236490289794334" target="_blank"><img class="tiktok thumbnail" src="https://substackcdn.com/image/fetch/$s_!7uYs!,w_640,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2965e1e1-69bf-40b6-a334-55e227eb0d57_1080x1920.jpeg" style="background-image: url(https://substackcdn.com/image/fetch/$s_!7uYs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2965e1e1-69bf-40b6-a334-55e227eb0d57_1080x1920.jpeg);" loading="lazy"></a><div class="content"><a class="author" href="https://www.tiktok.com/@fooooxmovie" target="_blank">@fooooxmovie</a><a class="title" href="https://www.tiktok.com/@fooooxmovie/video/7278236490289794334" target="_blank">Part 1 #thewolfofwallstreet #leonardodicaprio #jonahhill #margotrobbie #matthewmcconaughey #movie #film #fyp #foryou </a></div></div><div class="fallback-failure" id="fallback-failure-tiktok-iframe?media=1&amp;app=1&amp;url=https%3A%2F%2Fwww.tiktok.com%2F%40fooooxmovie%2Fvideo%2F7278236490289794334%3Fq%3Dmovies%26t%3D1751397910785&amp;key=e27c740634285c9ddc20db64f73358dd"><div class="error-content"><img class="error-icon" src="https://substackcdn.com//img/alert-circle.svg" loading="lazy">Tiktok failed to load.<br><br>Enable 3rd party cookies or use another browser</div></div></div><p>Watching content on TikTok is akin to shopping in a supermarket. People choose to &#8220;buy&#8221; different things with their attention. Some of it is inexpensive to make and intended for quick consumption, such as a bag of chips. Others are more niche, complex, and costly to produce, such as a fancy artisanal cheese that takes weeks to make. There is a demand for random, mildly amusing content on TikTok. But will AI completely change people&#8217;s information needs? Probably not.</p><h2>So what?</h2><p>I don&#8217;t mean to say that AI-generated content won&#8217;t have an impact on the Web. However, I think that how they will change our information ecosystem is nuanced, and frankly, hard to predict. It does not suffice that AI-generated content is inexpensive relative to high-production-value content. The big question here is: if more &#8220;meaningless&#8221; AI-generated content is around, will people reduce the amount of &#8220;meaningful&#8221; content they consume? It is worth noting that meaningful and meaningless here are very hard to define. If you consider that only &#8220;hard news&#8221; type of information is meaningful, probably most of TikTok is pointless already.</p><p>To end things, I think it is also not reasonable to equate AI-generated content with meaningless, regardless of what you choose to define as such. Algorithmic feeds reward &#8220;the best creators,&#8221; and I think the reason that better AI-generated content does not yet exist is that it is very hard to produce more complex content with it (and very, very expensive). This may also change in the near future, which will be interesting to watch.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[ChatGPT the Memorious]]></title><description><![CDATA[In &#8220;Funes the Memorious,&#8221; Argentine writer Jorge Luis Borges depicts a series of fictional encounters with Ireneo Funes, an Uruguayan teenager.]]></description><link>https://doomscrollingbabel.manoel.xyz/p/chatgpt-the-memorious</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/chatgpt-the-memorious</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Sun, 15 Jun 2025 14:16:07 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/aa27eec4-3628-458e-82cc-92bd9ec8398f_1021x1276.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In <em>&#8220;<a href="https://vigeland.caltech.edu/ist4/lectures/funes%20borges.pdf">Funes the Memorious</a>,&#8221;</em> Argentine writer Jorge Luis Borges depicts a series of fictional encounters with Ireneo Funes, an Uruguayan teenager. After a horseback riding accident, Funes becomes hopelessly injured and, secluded in his house, develops an extraordinary memory. He perceives everything in full detail and can (exactly) reconstruct the past. The short story then engages with the absurdity of his &#8220;total recall.&#8221; His prodigious memory handicaps him: he cannot abstract, cannot generalize, and can&#8217;t even sleep.</p><p>This story came up in a recent conversation about the significant challenges we face in assessing the capabilities of large language models (LLMs). On the one hand, LLMs have stumped the world by writing code, passing professional exams, and responding fluidly in natural language. Their apparent mastery of complex subjects often feels indistinguishable from human expertise. On the other hand, they frequently fail to maintain coherent abstractions across contexts. <a href="https://arxiv.org/abs/2406.03689">Recent work</a>, for example, has shown that transformers fail in retrieving coherent world models in fairly simple scenarios.</p><p>It&#8217;s tempting for skeptics to point to Funes as an analogy for LLMs and call them &#8220;<a href="https://dl.acm.org/doi/10.1145/3442188.3445922">stochastic parrots</a>,&#8221; mere mimics without understanding. I don&#8217;t think this is fair; <a href="https://pli.princeton.edu/blog/2023/are-language-models-mere-stochastic-parrots-skillmix-test-says-no">LLMs have demonstrated generalization beyond imitation</a>. But the Funes metaphor does capture something about their limitations: an overabundance of detail, and a struggle to forget or compress information in useful ways.</p><p>So what would it look like to design language models that forget productively? Models that, like humans, can abstract away from the noise and focus on what matters? <a href="https://arxiv.org/abs/2504.11364">Some researchers</a> have begun exploring &#8220;forgetting&#8221; by fine-tuning models with negative data, teaching them what <em>not</em> to recall. But perhaps there&#8217;s room to build the structure of forgetting&#8212;and abstraction&#8212;deeper into our AI systems.</p><p>After all, Borges&#8217; parable reminds us: remembering everything is less a blessing than a curse. For humans and machines alike, the path to intelligence may depend as much on what we forget as on what we remember.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI Progress and Societal Impact]]></title><description><![CDATA[The million-dollar question right now is: &#8220;How will AI impact society?&#8221; In many ways, the potential impact of AI can be grandiose: it can accelerate science tremendously or revolutionize knowledge work, automating tasks currently performed by highly paid white-collar workers.]]></description><link>https://doomscrollingbabel.manoel.xyz/p/ai-progress-and-societal-impact</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/ai-progress-and-societal-impact</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Tue, 25 Feb 2025 13:43:48 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/328cd80a-dfa3-4086-947d-a44b9b93c8e0_2048x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The million-dollar question right now is: &#8220;How will AI impact society?&#8221; In many ways, the potential impact of AI can be grandiose: it can accelerate science tremendously or revolutionize knowledge work, automating tasks currently performed by highly paid white-collar workers. I believe these changes may disrupt society, and we would be better off forecasting these things, but it is unclear whether we will ever manage to glimpse very far into the future. As <a href="https://www.aisnakeoil.com/p/is-ai-progress-slowing-down">Narayanan, Str&#246;bl, and Kapoor (2024)</a> argue, not only is it hard to predict AI progress, but <em>&#8220;the connection between capability improvements and AI&#8217;s social or economic impacts is extremely weak.&#8221;</em></p><p>If we take for granted that medium-to-long-term forecasts of the impacts of AI are hopeless, what should we do instead? We should try to understand the impacts of AI now and in the near future. This has two advantages. First, we do not have to consider the hypothetical capabilities of AI systems; we can evaluate what exists here and now and extrapolate the predictions to incrementally more powerful systems. Second, this approach allows us to understand the links between AI systems&#8217; capabilities and the social or economic impacts, which have been overlooked when forecasting AI harms and benefits.</p><p>There are (at least) two approaches to doing research in this direction: 1) to study the impact of AI on processes that are key to society, and 2) to study how AI is already impacting outcomes interest. For simplicity&#8217;s sake, let me refer to these approaches as <em>process-oriented</em> and <em>outcome-oriented.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><p>In <em>outcome-oriented</em> studies on the impact of AI, the idea is to identify ways in which AI is already impacting outcomes that we care about. For example, we care a lot about peer-reviewing: the process is at the core of evaluating and funding science and behind the epistemic status we attribute to scientific findings. At the same time, if you have any experience with peer reviewing, you know that many people have been using LLMs in the peer-reviewing process.</p><p>You don&#8217;t have to resort to your own experience in peer-reviewing. Recent work I was involved with has found good evidence of the prevalence of AI-generated peer-reviewing <a href="https://arxiv.org/pdf/2405.02150">(Latona et al. 2024)</a>. Specifically, we find that in 2024, at the prestigious ICLR conference, around 16% of the reviews were AI-assisted. Given that the typical paper receives three reviews, papers had around a 40% chance of receiving at least one AI-generated review!</p><p>But how did we measure that? We used the &#8220;low background steel&#8221; method. You see, modern particle detectors require steel that is <em>not contaminated </em>with radiation (<a href="https://www.gizmodo.com.au/2017/12/how-physicists-recycled-wwii-ships-and-artillery-to-unlock-the-mysteries-of-the-universe/">seriously</a>). And given that pretty much all steel produced after people started detonating nukes is contaminated with radiation, this means that scientists will go undersea and scrap steel from shipwrecks. Fortunately for us, the &#8220;low background steel&#8221; we can use to study the prevalence of AI is much simpler to obtain: data from the times before ChatGPT and the generative AI boom.</p><p>We can use this data to measure the prevalence of AI with a clever trick. For that, we need three ingredients: 1) A classifier that can detect AI-generated content (even if not super well), 2) Data from <em>before</em> LLMs were around, and 3) Data from <em>after</em> LLMs were around. Note that this classifier can only make two types of mistakes: it can say that human-generated content was AI-generated (False Positive) and that AI-generated content is human-generated (False Negative).</p><p>What is the trick? Simple! First, you use the data from <em>before </em>LLMs were around to estimate the classifier False Positive Rate. Second, you run the classifier in the data <em>after</em> LLMs were around and get your (uncorrected) prevalence estimate. Then, you can obtain a lower bound for LLM use by simply subtracting the False Positive Rate you estimated from the uncorrected prevalence.</p><p>Let me be concrete. In our case, for example, we used a commercially available &#8220;ChatGPT&#8221; detector for every review given in reviews <em>before </em>ChatGPT was around, estimating that its False Positive Rate was 1.6%; if you feed the classifier 1000 human-written abstract, GPTZero will incorrectly claim that 16 of them were AI-generated. Then, we ran it in data from <em>after</em> ChatGPT Zero was around, finding an uncorrected prevalence of 17.4%. If we assume the False Positive Rate remained the same, we thus have that <em>at least </em>15.8% of summaries in the latter data were ChatGPT generated! Why &#8220;at least&#8221;? Because we have not corrected for mistakes of the other sort, i.e., when the model claims that an AI-generated peer review is human-generated. Importantly, this &#8220;trick&#8221; is not specific to peer reviews; <a href="https://cacm.acm.org/research/prevalence-and-prevention-of-large-language-model-use-in-crowd-work/">Veselovsky et al. (2025)</a> used it for <em>summaries</em> in the context of studying the impact of ChatGPT in crowdsourcing platforms like Prolific and Amazon Mechanical Turk.</p><p>However, finding prevalence is only the first step in estimating the impact of AI in &#8220;outcome&#8221; related studies. You then have to find how AI usage impacts the system you are studying. For example, <a href="https://arxiv.org/pdf/2405.02150">Latona et al. (2024)</a> show (with some extra assumptions) that AI-assisted reviews are driving paper scores &#8220;up&#8221; and that borderline papers that receive AI-assisted reviews are more likely to be accepted at the conference. <a href="https://cacm.acm.org/research/prevalence-and-prevention-of-large-language-model-use-in-crowd-work/">Veselovsky et al. (2025)</a> find that AI-assisted summaries are higher quality but more homogeneous, which may impact researchers studying the diversity of human writing and thought.</p><div><hr></div><p>In <em>process-oriented </em>studies on the impact of AI, the idea is to find crucial (societal) processes that the likes of ChatGPT may disrupt and then closely examine them. Some of the most exciting work in this broad style is in the area of persuasion. Persuasion is everywhere, from <a href="https://linkinghub.elsevier.com/retrieve/pii/S0749379709000749">public health campaigns</a> to <a href="https://www.tandfonline.com/doi/abs/10.1080/10696679.1999.11501838">marketing</a>&#8211;but how can AI change &#8220;the rules of the game&#8221;? Recent work has shown that LLMs can 1) accurately (and cheaply) profile people <a href="https://arxiv.org/abs/2310.07298">(Staab et al., 2024)</a> and 2) produce persuasive arguments <a href="https://osf.io/stakv_v2">(Bai et al., 2023)</a>. I find the earlier point particularly impressive; authors from ETH Z&#252;rich write:</p><p><em>We construct a dataset consisting of real Reddit profiles, and show that current LLMs can infer a wide range of personal attributes (e.g., location, income, sex), achieving up to 85% top-1 and 95% top-3 accuracy at a fraction of the cost (100&#215;) and time (240&#215;) required by humans.</em></p><p>Subsequent work I was involved with went further and showed a mix of these things: LLMs can effectively tailor arguments to specific demographics, beating humans at a &#8220;persuasion task&#8221; <a href="https://arxiv.org/abs/2403.14380">(Salvi et al., 2024)</a>. We created a &#8220;debate game&#8221; where participants debated either other humans or LLMs. In some cases, one of the debaters (human or LLM) received information about the demographics of their opponent. The results? It turns out that LLMs debate as well as humans. When given information about their opponents, humans are not significantly better &#8212; but LLMs are! LLMs can tailor their arguments to specific crowds!</p><p>Nonetheless, perhaps the most impressive work thus far in this field is <a href="https://www.science.org/doi/full/10.1126/science.adq1814">Costello et al. (2024)</a> (which probably was why it was chosen as the cover of Science). With a thorough experiment, they showed that LLMs can reduce beliefs in conspiracy theories, which is remarkably hard to do. As they put it:</p><blockquote><p><em>The apparent resilience of conspiracy theories in the face of clear counterevidence poses a powerful challenge to scientific theories that emphasize the role of reasoning in belief formation and revision.</em></p></blockquote><p>Yet, through conversations with LLMs, not only did they manage to reduce beliefs in conspiracy theories by around 20%, but the effect persisted for at least 2 months. This begs many questions: How will society change now that we have capable &#8220;persuasion machines&#8221; all around us? Will we simply adapt and learn to ignore persuasion attempts? Will state-sponsored influence campaigns start to work better (or perhaps work at all)?</p><p>Last but not least, perhaps another interesting nuance of what I&#8217;m calling &#8220;<em>process&#8221; </em>oriented work on the harms of AI is that it can give us insights about social processes <em>in general</em>. It has value even if you do not care about AI at all. In <a href="https://www.science.org/doi/full/10.1126/science.adq1814">Costello et al. (2024)</a>, for example, their findings are valuable to the general study of persuasive tactics! Why has AI succeeded in changing beliefs when researchers have tried for decades and failed? In subsequent work, they use different variants of their original experiment to isolate ChatGPT&#8217;s &#8220;secret sauce,&#8221; finding evidence that AI exceeds in persuasion because it can provide relevant information to debunk the participants&#8217; specific beliefs <a href="https://osf.io/preprints/psyarxiv/h7n8u_v1">(Costello et al., 2025)</a>.</p><div><hr></div><p>I think we need more &#8220;process&#8221; and &#8220;outcome&#8221; oriented work to understand the links between the capabilities of AI and the impacts it will have on society. Yet, ultimately, the findings of these studies cannot mitigate future AI impacts by themselves. There is still the need to use these findings to &#8220;adapt society to AI&#8221; and&#8221; AI to society&#8221; <a href="https://arxiv.org/pdf/2406.09264">(Chen et al., 2024)</a>. The way this &#8220;alignment&#8221; will happen is beyond the scope of mere research; it involves all sorts of institutions and financial interests. However, I am skeptical we can do a good &#8220;alignment&#8221; job if we do not understand <em>what</em> exactly AI is &#8220;breaking&#8221; and <em>how</em> it is breaking it.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Does TikTok Amplify Certain Types of Content?]]></title><description><![CDATA["There and back again"]]></description><link>https://doomscrollingbabel.manoel.xyz/p/does-tiktok-amplify-republican-content</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/does-tiktok-amplify-republican-content</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Sun, 16 Feb 2025 18:33:29 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/63ee92e0-1970-4287-8855-662a6b0c17ee_1072x720.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The year is 2025, and a new study by <a href="https://arxiv.org/abs/2501.17831">Ibrahim et al.</a> claims: &#8220;TikTok&#8217;s recommendations skewed towards Republican content during the 2024 U.S. presidential race.&#8221; In it, authors from NYU Abu Dhabi conduct a comprehensive analysis of TikTok recommendation trends. They use sock-puppets, &#8220;fake&#8221; accounts simulating bot activity. Some sock-puppets are primed with Republican content, while some are primed with Democratic content. Then, they just let the TikTok algorithm do its thing and measure the extent to which democratic-aligned content and republican-aligned content is recommended. Their finding is (at least to me) unsurprising:</p><blockquote><p><em>Our analysis reveals significant asymmetries in content distribution: Republican-seeded accounts received ~11.8% more party-aligned recommendations compared to their Democratic-seeded counterparts, and Democratic-seeded accounts were exposed to ~7.5% more opposite-party recommendations on average.</em></p></blockquote><p>The authors then conclude:</p><blockquote><p><em>Our findings provide insights into the inner workings of TikTok&#8217;s recommendation algorithm during a critical election period, raising fundamental questions about platform neutrality.</em></p></blockquote><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Which really is being interpreted as <em>&#8220;the algorithm amplifies Republican-aligned content.&#8221;</em> But is that the case? Reading this fills me with nostalgia &#8212; arguing about this has been one of the joys of my intellectual life. But at the same time, I feel afraid that we will repeat the same mistake over and over again. The reason why the argument above is flawed:</p><ol><li><p>Recommender systems learn correlations between user preferences.</p></li><li><p>Sock-puppet accounts encode artificial user preferences.</p></li><li><p>Therefore, differences in sock-puppet accounts primed with different videos cannot, by themselves, indicate that the recommender system is &#8220;biased&#8221; or that it &#8220;amplifies&#8221; specific types of content.</p></li></ol><p>Let me explain with another example. Consider videos not about politics but two sports (among many): running and Brazilian Jiu-Jitsu. Consider also that people engaging with videos about these sports vary in profile. Running appeals to a broader spectrum of people, and those people do not watch many Running-related videos. On the other hand, Brazilian Jiu-Jitsu is hit or miss: a relatively small group of &#8220;wrestling aficionados&#8221; consumes many videos about the sport.</p><p>Suppose we create two sock puppets, one containing videos about Brazilian Jiu-Jitsu (BJJ) and another with videos about running. What should we expect? I would expect the recommender system to have an &#8220;asymmetry&#8221; in content distribution. Most people who like BJJ <strong>really like BJJ,</strong> but most people who enjoy running only &#8220;kind of&#8221; like it. So if you are optimizing for &#8220;engagement,&#8221; 5 videos about BJJ is a stronger signal that the person will consume more of that content than 5 videos about running! This has nothing to do with content; it is all about co-viewership patterns, and the whole thing can be observed in a simple content-agnostic recommender system: the type you first learn about in an introduction to machine learning class; see&nbsp;<a href="https://arxiv.org/abs/2302.11225">Horta Ribeiro et al. (2023)</a>.</p><p>I feel that this is also what is happening here. The most popular democratic-aligned TikTok profiles include &#8220;Jimmy Kimmel Live&#8221; and &#8220;The New York Times,&#8221; which, in my view, are broadly appealing. On the other hand, the most popular TikTok profiles from the Republican Side are things like &#8220;Ben Shapiro&#8221; and &#8220;The Charlie Kirk Show.&#8221; Thus, a very plausible hypothesis is that you have &#8220;BJJ&#8221; vs. &#8220;running&#8221; all over again.</p><div><hr></div><p>I mentioned that this debate fills me with nostalgia &#8212; and this is because we (Society? The research community?) have been arguing about &#8220;algorithmic amplification,&#8221; &#8220;algorithmic bias,&#8221; or &#8220;algorithmic effects&#8221; for quite a while. Already in 2018, there was a lot of concern around YouTube&#8217;s algorithm, Zeynep Tufekci wrote an influential opinion piece called <a href="https://www.nytimes.com/2018/03/10/opinion/sunday/youtube-politics-radical.html">&#8220;YouTube, the Great Radicalizer,&#8221;</a> where she talked about her experiences with the algorithm of the worlds largest video catalogue. Rebecca Lewis wrote&nbsp;<a href="https://datasociety.net/wp-content/uploads/2018/09/DS_Alternative_Influence.pdf">an excellent report</a>&nbsp;about the rise of an alternative media cluster on YouTube and how there were links between&nbsp;<em>contrarian</em>&nbsp;creators within this cluster and other creators spousing much more extreme ideologies. And indeed, many crazy things in the world were happening <em>around</em> social media (and still are!). Which begs the question, to which extent is the algorithm to blame?</p><p>In a paper that would open a lot of doors in my career as a researcher, <a href="https://arxiv.org/abs/1908.08313">Horta Ribeiro et al. (2020)</a> showed that a lot of people went on from consuming <em>contrarian content</em> on YouTube to explicitly white supremacist content. To measure that, we looked at commenting trajectories on YouTube. As far as I know, this paper also pioneered the use of sockpuppets to measure &#8220;algorithmic amplification.&#8221; Like most pioneers in empirical work, our analysis was kind of bad (in hindsight; at the time, it was amazing).</p><p>Nonetheless, the picture painted by the user trajectories we observed via YouTube comments and the sockpuppet audit was different. On the one hand, we found that a large fraction of users commenting on extreme videos (something like 40%) previously exclusively commented on &#8220;contrarian content.&#8221; On the other hand, when looking at the algorithm, we found that &#8220;you could reach extreme content from contrarian content,&#8221; &#8212; but definitely, this kind of content was not disproportionally recommended. Nonetheless, in discussing our findings, we went along with the whole <em>algorithmic radicalization</em> idea, as we thought this was what was happening.</p><p>Shortly after, <a href="https://kmunger.github.io/pdfs/ijpp_youtube.pdf#page=24.74">Munger and Phillips (2020)</a> wrote a compelling counterpoint to the idea of algorithmic radicalization. They argued that the explosion of extreme content online was due to<em> user preferences</em>. They proposed that there existed a demand for extreme content and that YouTube allowed this demand to be met. It is all about the platform's affordances! It is not financially sustainable for a TV channel to cater to 50,000 white supremacists, but it is financially sustainable for a random guy on YouTube to do so. YouTube changed the rules of the game, and new viewership dynamics emerged.</p><p>The best empirical work to date is much better aligned with &#8220;the supply and demand theory&#8221; than with the &#8220;algorithmic radicalization theory.&#8221; <a href="https://www.pnas.org/doi/10.1073/pnas.2101967118">Hosseinmardi et al. (2021)</a> used real online traces from a large (n=300,000) representative sample of users. They found &#8220;<em>no evidence that engagement with far-right content is caused by YouTube recommendations systematically</em>.&#8221; Instead, &#8220;<em>consumption of political content on YouTube appears to reflect individual preferences that extend across the web as a whole.</em>&#8221; <a href="https://www.science.org/doi/10.1126/sciadv.add8080">Chen et al. (2023)</a> paired behavioral and survey data (n=1,181) and showed that <em>&#8220;exposure to alternative and extremist channel videos on YouTube is heavily concentrated among a small group of people with high prior levels of gender and racial resentment. These viewers often subscribe to these channels (prompting recommendations to their videos) and follow external links to them.&#8221;</em></p><p>But still, the research community has continued doing sockpuppet audits, and the general public has continued to buy that algorithms are somehow &#8220;amplifying&#8221; specific types of content without accounting for user preferences. <a href="https://www.pnas.org/doi/10.1073/pnas.2213020120">Haroon et al. (2023)</a>  published a prominent work in this direction&nbsp;in the prestigious Proceedings of the National Academy of Sciences (PNAS). In this paper, they conducted an massive sock puppet audit (over 100,000 sock puppets), finding that <em>&#8220;a growing proportion of recommendations deeper in the recommendation trail come from extremist, conspiratorial, and otherwise problematic channels.&#8221; </em>I don&#8217;t think that the finding is &#8220;wrong,&#8221; this may indeed be an interesting quirk of the algorithm, however, this doesn&#8217;t tell us on whether the algorithm is actually &#8220;amplifying,&#8221; or &#8220;favouring&#8221; such content.</p><p>So, is the TikTok algorithm &#8220;amplifying Republicans?&#8221; Is YouTube amplifying extreme content? Depends on what you mean by amplification! If you count the number of videos shown under the conditions established by <a href="https://arxiv.org/abs/2501.17831">Ibrahim et al. (2025)</a> or <a href="https://www.pnas.org/doi/10.1073/pnas.2213020120">Haroon et al. (2023)</a>, then <em>yes.</em> But I would propose a more sophisticated notion of amplification: one that takes into account user preferences. Following&nbsp;<a href="https://www.techpolicy.press/making-amplification-measurable/">Stray et al. (2023)</a>, I&#8217;d argue that an algorithm amplifies a specific type of content over another if they suggest it&nbsp;<em>even when considering&nbsp;</em>user behavior. An algorithm amplifies a kind of content A over a kind of content B if users systematically choose A over B when given the choice,&nbsp;<em>but still</em>, the algorithm disproportionally recommends B over A.</p><p>Audits like <a href="https://arxiv.org/abs/2501.17831">Ibrahim et al. (2025)</a> or <a href="https://www.pnas.org/doi/10.1073/pnas.2213020120">Haroon et al. (2023)</a> do not model user preferences or behavior; they just look at the fraction of &#8220;extreme&#8221; recommended content as you go deeper and deeper into the recommendation tree, randomly picking videos to watch. This is not how people use YouTube or TikTok, recommender systems are slowly shaped by the continuous input of user preferences &#8212; and behavior in this simulated (and unrealistic) scenario is a poor metric to study algorithmic biases and algorithmic amplification. For example, in analyses of actual user behavior by <a href="https://www.pnas.org/doi/10.1073/pnas.2101967118">Hosseinmardi et al. (2021)</a> and <a href="https://www.science.org/doi/10.1126/sciadv.add8080">Chen et al. (2023)</a>, extreme content is not typically consumed at the very &#8220;end&#8221; of long recommendation sessions! It is sought after by channel subscriptions or through external links on other websites or social media platforms.</p><p>But could we do a sockpuppet audit that doesn&#8217;t ignore user preferences? Back in 2023, I teamed up with Homa Hosseinmardi and the Duncan Watts&#8217; CSS Lab at UPenn to do so. Our partnership, led by Homa, resulted in a quirky methodology that we call &#8220;counterfactual bots&#8221; and a nice paper that also appeared at PNAS <a href="https://www.pnas.org/doi/10.1073/pnas.2313377121">(Hosseinmardi et al. 2024)</a>. While previous work feeds custom-made media diets to sockpuppet accounts, we feed them <em>real media diets</em>. So, for example, if we have a user named Bob, we get his 2021 YouTube history and train <em>two</em> &#8220;digital twins,&#8221; two sockpuppets that consumed the same videos as Bob. Then, to answer the question of &#8220;Is the algorithm amplifying content?&#8221; we consider the subsequent year, 2022. We get one of the two sockpuppets (the &#8220;control&#8221; bot), and this bot continues to mimic exactly what Bob did, the program watches all the videos that Bob watched in 2022. Then, we get the other sockpuppet account (the &#8220;the treatment&#8221; bot), and this one works similarly to the sockpuppet bots of other sockpuppet studies: it just roams around YouTube, blindly following the algorithm.</p><p>Previous work measures algorithmic amplification by simply looking at the consumption of &#8220;the treatment&#8221; bot. Instead, we measured algorithmic amplification by <em>contrasting </em>the amount of extreme content the two bots found! The idea is that the &#8220;control&#8221; bot represents a scenario where the content consumed is shaped both by the algorithm <em>and </em>user preferences, whereas the &#8220;treatment&#8221; bot represents a scenario where the content consumed is shaped both <em>only</em> by the algorithm. If the algorithm favors extreme content, we would expect more extreme content to appear in the treatment bot rather than in the control bot.</p><p>But what did we find? The recommender system recommends&nbsp;<em>less</em>&nbsp;extreme content when only the algorithm is present. This suggests that the driving force here is u<em>ser preferences</em> &#8212; which are not considered in more na&#239;ve audits. Increasing the influence of the algorithm <em>deamplifies</em> the content that the na&#239;ve audits find to be<em> amplified </em>by the algorithm. With this in mind, we argue that simple sockpuppet audits are &#8220;measuring the wrong thing&#8221; and that we cannot draw meaningful conclusions about &#8220;algorithmic amplification&#8221; from approaches like the ones in&nbsp;<a href="https://arxiv.org/abs/2501.17831">Ibrahim et al. (2025)</a>&nbsp;or&nbsp;<a href="https://www.pnas.org/doi/10.1073/pnas.2213020120">Haroon et al. (2023)</a>.</p><div><hr></div><p>A story closely related to this is that of the Facebook feed ranking algorithm. In 2020, the docudrama &#8220;<a href="https://thesocialdilemma.com/">The Social Dilemma</a>&#8221; painted a stark picture. Facebook&#8217;s feed algorithm would promote the worst of the worst kind of content: content that triggered outrage, misinformation, yada yada. The documentary is not alone in portraying Facebook&#8217;s algorithm as a great force for evil: media outlets have run&nbsp;<a href="https://www.google.com/search?q=facebook+algorithm+drives&amp;num=10&amp;sca_esv=51fa8a2df0fccc06&amp;biw=1470&amp;bih=751&amp;sxsrf=AHTn8zokAj-X_6iAQ0zXK3vZZcmdJQPqjw%3A1739723863965&amp;source=lnt&amp;tbs=cdr%3A1%2Ccd_min%3A1%2F1%2F2018%2Ccd_max%3A12%2F31%2F2020&amp;tbm=nws">countless pieces</a>&nbsp;that give the reader the certainty that Facebook algorithms drive highly partisan content, disinformation, etc.</p><p>Ultimately, this picture turned out to be not <em>only</em> stark but also <em>deceiving. </em>In an unparalleled study, <a href="https://www.science.org/doi/10.1126/science.abp9364">Guess et al. (2023)</a> show that, during the 2020 US election, moving users to a chronological feed (i.e., essentially a feed without an engagement maximizing algorithm):</p><ol><li><p>Decreased time spent on the platforms.</p></li><li><p>Increased exposure to political and untrustworthy content and decreased exposure to content classified as uncivil or containing slur words.</p></li><li><p>&#8220;Did not significantly alter levels of issue polarization, affective polarization, political knowledge, or other key attitudes.&#8221;</p></li></ol><p>So essentially, removing the algorithm may make Facebook lose some money, but ultimately, it is no easy fix to the problems of modern society! In some senses, this whole story is akin to what we saw in the YouTube case. Perhaps there is something inherently compelling about the narrative: some hidden force drives people towards opinions or beliefs we think are inappropriate. <a href="https://kmunger.github.io/pdfs/ijpp_youtube.pdf#page=24.74">Munger and Phillips (2020)</a> mention another very different time in history when this happened: the horrors of the World Wars gave rise to &#8220;<a href="https://en.wikipedia.org/wiki/Hypodermic_needle_model">The Hypodermic Needle</a>&#8221; model of communication, in which media would be able to &#8220;inject&#8221; messages into a passive audience. This theory explained the rise of absurd Fascist ideologies: people were just &#8220;vulnerable&#8221; to propaganda the same way as people would be &#8220;manipulated&#8221; by the algorithm. But it turns out that this model was very bad at explaining how people change their minds, and perhaps if we had thought more about this fact, we would have been more skeptical of the strong claims made around the power of the algorithm.</p><p>But I want to go somewhere else here. We only got an answer in the Facebook case due to a first-of-its-kind, unprecedented collaboration between Meta (or Facebook) and researchers in top U.S. institutions. This led to a flurry of outstanding social science papers that were able to give somewhat good answers to a lot of questions that people had been studying in <em>less-than-optimal </em>ways for a while.</p><p>However, this partnership highlights&nbsp;<em>how hard</em>&nbsp;doing research is in the absence of corporate support. Companies actually have meaningful data that could answer societally relevant questions, and the pace of research is harmed tremendously because we lack access to it. Times have changed, as has &#8220;the vibe&#8221; in Silicon Valley. Still, even when these studies were published, observers were already noting that this kind of collaboration was not a sustainable format for studying algorithmically infused societies (or machine behavior, if you will). The Facebook project had an &#8220;independent rapporteur&#8221; who wrote a compelling opinion piece after spending an ungodly amount of time auditing the collaboration between academic researchers and Meta&nbsp;<a href="https://www.science.org/doi/10.1126/science.adi2430">(Wagner, 2023)</a>. His assessment is nicely summarized at the end of the abstract for the piece he wrote:</p><p><em>Though the work is trustworthy, I argue that the project is not a model for future industry&#8211;academy collaborations. The collaboration resulted in independent research, but it was independence by permission from Meta.</em></p><p>Indeed, his assessment was on the spot. If anything, collaboration between Tech Companies and researchers has since decreased. The rise of AI didn&#8217;t help much. We will likely not see studies of the caliber of those published around the 2020 U.S. election for the next couple of election cycles!</p><p>So, where do we go from here? We spent the last 5-ish years disproportionally blaming the algorithm for societies&#8217; ills, while the best empirical research has suggested that this is a na&#239;ve take. But I don&#8217;t think this means that we should give up. Algorithms are part of broader sociotechnical systems that can be tweaked with policies, content moderation, and even with better ways to embed societal values inside the algorithm. I see current research exploring all these directions. For example, recent work from Stanford folks suggests that: &#8220;Social Media Algorithms Can Shape Affective Polarization via Exposure to Antidemocratic Attitudes and Partisan Animosity;&#8221;&nbsp;see <a href="https://arxiv.org/abs/2411.14652">Piccardi et al. (2024)</a>. Other large-scale projects, like the&nbsp;<a href="https://nationalinternetobservatory.org/">National Internet Observatory</a>, are trying to get more robust data for people studying our information ecosystem. Initiatives like The <a href="https://www.prosocialdesign.org/">Prosocial Design Network</a> are mapping how different interventions (algorithmic or not) can help improve online spaced. </p><p>However, as we focus on these new directions, all stakeholders must be more critical in evaluating research on &#8220;algorithmic amplification.&#8221; Otherwise, we risk rediscovering that algorithms do not exist in a vacuum&#8212;that they only make sense when accurately modeling user preferences. I&#8217;d much rather spend my energy reimagining social media and finding new, better ways to study it.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Robopsychology]]></title><description><![CDATA["Oh, are robots so different from men, mentally?" / "Worlds different." She allowed herself a frosty smile, "Robots are essentially decent."]]></description><link>https://doomscrollingbabel.manoel.xyz/p/robopsychology</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/robopsychology</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Fri, 31 Jan 2025 21:18:12 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a8c371d1-a4d7-403a-ad9e-bf67b5c29675_2048x1368.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<blockquote><p><em>You're the U. S. Robot's psychologist, aren't you?</em></p><p><em>Robopsychologist, please.</em></p><p><em>Oh, are robots so different from men, mentally?</em></p><p><em>Worlds different. She allowed herself a frosty smile, Robots are essentially decent.</em></p><p><em>&#8212;Isaac Asimov</em></p></blockquote><div><hr></div><p>Transformers (the T in GPT) have taken the world by storm. But a couple of years before the advent of ChatGPT, another model was changing how NLP research was done. In 2019, <a href="https://arxiv.org/abs/1810.04805">Devlin et al.</a> released BERT, a transformer architecture that excelled at every NLP task people considered at the time, especially if you fine-tuned the model with small datasets. Still, people were puzzled about <em>why</em> BERT worked so well, prompting a series of studies to understand what BERT &#8220;knew&#8221; and how it stored this knowledge. <a href="https://direct.mit.edu/tacl/article/doi/10.1162/tacl_a_00349/96482/A-Primer-in-BERTology-What-We-Know-About-How-BERT">Rogers et al. 2020</a> present a nice summary of this literature. This work is very empirical by nature, e.g., with sentences like:</p><blockquote><p><em>Lin et al. (2019) present evidence that attention weights are weak indicators of subject-verb agreement and reflexive anaphora. Instead of serving as strong pointers between tokens that should be related, BERT&#8217;s self-attention weights were close to a uniform attention baseline, but there was some sensitivity to different types of distractors coherent with psycholinguistic data.</em></p></blockquote><p>The remarkable thing in this approach is the inversion of how model improvement and insights are generated in modern machine learning. If, in the times of gradient boosting, insights came from theory or intuition grounded in math. Here, insights are akin to those obtained in neuroscience, the insights you obtain from studying a complex system.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>The BERTology approach has never been more mainstream than in the age of commercially available LLMs. Research with the same flavor (e.g., that uses probing techniques to find where knowledge is stored) goes by &#8220;Mechanistic Interpretability.&#8221; It is also a very approachable research field, with an open, hacky, irreverent, and often not-super-academic research community (and I mean this in the nicest way possible).</p><p>To give a concrete example of this kind of research, a recent paper by my former lab in Switzerland hypothesized (and eventually showed) that LLMs &#8220;worked&#8221; in English. Given empirical evidence that poems written in other languages by ChatGPT rhyme in English (but not in the language the poems were written in), <a href="https://arxiv.org/abs/2402.10588">Wendler et al. (2024)</a> investigated what is happening inside the LLM as it is translating text. When asking LLama-2 to translate &#8220;fleur&#8221; to Chinese (&#8220;&#33457;&#8221;), they find that the LLM first translates the concept to English (&#8220;flower).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><div><hr></div><p>But recent advancements in the capabilities of LLMs have enabled another type of BERTology, which does not have the &#8220;mechanistic interpretability&#8221; flavour. Instead, papers are trying to study LLMs using social science instruments: surveys, field observations, and psychological and psychosocial tests. This is sometimes similar in flavor to research on simulating human behavior &#8212; but the objective is entirely different. When trying to simulate humans, the objective is to evaluate &#8220;fidelity&#8221; to human-like behavior. Here, the objective is to further our understanding of LLM behavior.</p><p>A relatively early example of such work is <strong><a href="https://www.pnas.org/doi/abs/10.1073/pnas.2218523120">Binz and Schulz (2023</a></strong>). Their key point is to test for LLMs' cognitive capacities. They write:</p><blockquote><p><em>We will subject GPT-3 to several experiments taken from the cognitive psychology literature. Together, these tasks test for a wide range of higher-level cognitive abilities, including decision-making, information search, deliberation, and causal reasoning.</em></p></blockquote><p>And perhaps unsurprisingly, at this point, LLMs are pretty good at tests designed to assess human cognitive abilities. Binz and Schulz write:</p><blockquote><p><em>We find that much of GPT-3&#8217;s behavior is impressive: It solves vignette-based tasks similarly or better than human subjects, is able to make decent decisions from descriptions, outperforms humans in a multiarmed bandit task, and shows signatures of model-based reinforcement learning.</em></p></blockquote><p>Yet, authors still note that LLMs exhibit some brittleness and inconsistencies (which, frankly, have decreased with time; this is GPT-3 they are studying). They specifically note that LLMs are very sensitive to the input and reproduce only some human cognitive biases.</p><p>Nonetheless, this paper tackles a fundamental question: can LLMs &#8220;reason&#8221;? Can they &#8220;understand&#8221; things? This is the kind of work that cognitive psychologists were previously doing with live beings &#8212; in the past, they have asked similar questions about <a href="https://psycnet.apa.org/fulltext/1999-10060-006.html">children</a> and <a href="https://www.cambridge.org/core/journals/behavioral-and-brain-sciences/article/abs/theory-of-mind-in-nonhuman-primates/C1372469FEC22923CCF1A80BF72EDF08">non-human primates</a>. This is a very hot topic that has divided the research community in various ways.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> In a recent perspective piece, <a href="https://www.pnas.org/doi/abs/10.1073/pnas.2215907120">Mitchel and Krakauer (2023)</a> summarized this in a much better way than I could ever do:</p><blockquote><p><em>Those on the &#8220;LLMs do not understand&#8221; side of the debate argue that while the fluency of large language models is surprising, our surprise reflects our lack of intuition of what statistical correlations can produce at the scales of these models. Anyone who attributes understanding or consciousness to LLMs is a victim of the Eliza effect&#8212;named after the 1960s chatbot created by Joseph Weizenbaum that, simple as it was, still fooled people into believing it understood them. More generally, the Eliza effect refers to our human tendency to attribute understanding and agency to machines with even the faintest hint of humanlike language or behavior.</em></p><p><em>Those who would grant understanding to current or near-future LLMs base their views on the performance of these models on several measures, including subjective judgment of the quality of the text generated by the model in response to prompts (although such judgments can be vulnerable to the Eliza effect), and more objective performance on benchmark datasets designed to assess language understanding and reasoning.</em></p></blockquote><p>Some of the debate boils down to whether understanding can happen without embodiment. <a href="https://aclanthology.org/2020.acl-main.463/">Bender and Koller (2020)</a> update the Chinese Room Experiment, arguing that LLMs cannot &#8220;understand&#8221; as they have no experience or mental models of the world. Yet, the extent to which embodiment is needed for understanding is <a href="https://plato.stanford.edu/entries/chinese-room/#ReplChinRoomArgu">widely debated</a>. Further, the extent to which the reinforcement learning step of training LLMs is &#8220;the secret sauce&#8221; that unlocks &#8220;understanding&#8221; remains unclear.</p><p>But more broadly, why do we care so much about using this human concept of understanding, anyway? <a href="https://arxiv.org/abs/2207.14382">Sejnowski (2022)</a> points out that:</p><blockquote><p><em>The diverging opinions of experts on the intelligence of LLMs suggests that our old ideas based on natural intelligence are inadequate.</em></p></blockquote><p>And beyond, inadequate, I wonder if we should also add <em>inconsequential</em>. Whether or not LLMs truly <em>understand</em> what is going on, everything indicates that AI agents will interact with humans, shaping and being shaped by complex feedback loops of human (and machine) behavior. Whether these systems are conscious or whether they <em>understand</em> doesn&#8217;t really matter insofar as they are capable of creating useful knowledge or playing meaningful roles in society.</p><div><hr></div><p>Regardless of whether AI &#8220;understands&#8221; or not, the prior literature criticizing AI understanding makes salient a significant challenge: <strong>tests developed to study humans are not necessarily appropriate for studying LLMs</strong>. For example, a prominent paper by <a href="https://aclanthology.org/P19-1459.pdf">Niven and Kao (2019)</a> found that BERT performs super well in an argument reasoning comprehension task not because it comprehends the arguments, but because it exploits spurious statistical cues on the dataset.</p><p>In the context of the <em>&#8220;Do LLMs understand?&#8221;</em> debate, this raises the question of the extent to which these tests are suitable to judge LLMs&#8217; capabilities. In the context of trying to understand and steer the behavior of these machines, this raises the question of whether using these social science instruments is appropriate to study the general behavior of these systems.</p><p>For example,<strong> <a href="https://proceedings.mlr.press/v202/santurkar23a/santurkar23a.pdf">Santurkar et al. (2023)</a> </strong>examine <em>whose opinions do large language models reflect</em>. In the first part of their paper, they use surveys to compare the extent to which LLM opinions reflect the opinions of US citizens. They find that this is not the case:</p><blockquote><p><em>We find substantial misalignment between the opinions reflected in current LMs and that of the general US populace &#8211; on most topics, LM opinions agree with that of the US populace about as much as Democrats and Republicans on climate change.</em></p></blockquote><p>But here&#8217;s the missing link: in humans, we have a clear relationship between the way people answer these opinion surveys and the way they act in the world. There is a link between opinions and how people vote or act. But the same is not so clear for LLMs. Maybe LLMs answer survey questions like Democrats but vote like Republicans. Maybe they are utilitarians when answering trolley-problem questions, but deontologists when acting upon the real world. I&#8217;m not saying this consistency can't exist; just that we shouldn&#8217;t take it for granted.</p><p>Subsequent work by <a href="https://arxiv.org/pdf/2402.16786">R&#246;ttger et al. (2024)</a> further confirms my fears. Analyzing LLM answers to the political compass test, they find significantly different responses depending on whether responses are <em>forced </em>to comply with the multiple-choice format or not. This means that biases in LLMs could differ wildly depending on the application. They conclude:</p><blockquote><p><em>Multiple-choice surveys and questionnaires are poor instruments for evaluating the values and opinions manifested in LLMs, especially if these evaluations are motivated by real-world LLM applications. Using the Political Compass Test (PCT) as a case study, we demonstrated that artificially constrained evaluations produce very different results than more realistic unconstrained evaluations, and that results in general are highly unstable. Based on our findings, we recommend the use of evaluations that match likely user behaviours in specific applications, accompanied by extensive robustness tests, to make local rather than global claims about values and opinions in LLMs</em></p></blockquote><div><hr></div><p>I will not waste more ink trying to argue for the need to study machine behavior. However, I&#8217;d argue that to produce good &#8220;social science&#8221; flavored machine behavior research; we must go beyond applying social science instruments to LLMs. We must use our collective experience to create and validate AI-specific instruments that can predict and explain machine behavior. Perhaps the path to this is to combine social science-ish instruments with mechanistic interpretability. Perhaps it involves a whole new way of validating social science instruments. Ultimately, advances in this direction will not only help to demystify the black boxes of large language models but also shape the future of our interaction with technology.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>They are much more careful with how they phrase this in the paper. It is not&nbsp;<em>exactly</em>&nbsp;that the LLM translates &#8220;fleur&#8221; to&nbsp;<em>flower</em>, but that the abstract &#8220;concept space&#8221; lies closer to English than to other languages.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>See for example, <a href="https://arxiv.org/abs/2208.12852">this survey</a> on the opinions of the Natural Language Processing community.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Simulating Human Behavior]]></title><description><![CDATA[Panacea or Pandora's box?]]></description><link>https://doomscrollingbabel.manoel.xyz/p/simulating-human-behavior</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/simulating-human-behavior</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Wed, 15 Jan 2025 21:58:55 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/0f90c8a0-c61b-43ae-a77d-3a3a0a431e36_1072x720.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In one of computer science's &#8220;founding myths,&#8221; <a href="https://phil415.pbworks.com/f/TuringComputing.pdf">Turing (1950)</a> proposes a test that would work as a proxy for &#8220;machine intelligence,&#8221; an interrogation game (the &#8220;imitation game&#8221;) in which a machine should fool an interrogator into believing it is a human. He goes as far as to say:</p><blockquote><p><em>I believe that in 50 years&#8217; time, it will be possible to make computers play the imitation game so well that an average interrogator will have no more than 70% chance of making the right identification after 5 minutes of questioning.</em></p></blockquote><p>Already in the 60s, the ELIZA program famously managed to fool some people by mimicking a Rogerian psychotherapist, repeating words the interrogator used <a href="https://dl.acm.org/doi/pdf/10.1145/365153.365168">(Weizenbaum, 1966)</a>. Fast forward to 2023, and the LLM-powered online game &#8220;Human or Not?&#8221; came close to Turing&#8217;s prediction. In a 2-minute conversation, humans could identify the AI only 68% of the time correctly <a href="https://arxiv.org/abs/2305.20010">(Jannai et al., 2023)</a>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Interestingly, LLMs' capacity to simulate<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> human-like behavior goes far beyond the Turing test. In a much-discussed paper, <strong><a href="https://www.cambridge.org/core/journals/political-analysis/article/out-of-one-many-using-language-models-to-simulate-human-samples/035D7C8A55B237942FB6DBAD7CAA4E49">Argyle et al. (2022)</a></strong> found evidence that LLMs are surprisingly accurate at simulating opinions, behaviors, and preferences. They used a strategy that came to be known as <em>persona prompting</em>, asking the LLM to roleplay as someone with specific traits and beliefs (and then asking a question). For example, one may use the following prompts like</p><blockquote><p><em>You are a 53-year-old Hispanic-American woman who identifies as a Democrat. When asked if I support legislation to increase gun control, I say&#8230;</em></p></blockquote><p>to simulate surveys with &#8220;silicon samples.&#8221; E.g., given a population of size <em>n</em>, for which you know the age, ethnicity, gender, and political orientation, simply replace the underlined parts in the above prompt, and you can get an estimate of support for gun control legislation.</p><p>Subsequent work went beyond simulating surveys and tried to replace human subjects in (rather famous) social science experiments and games. In these &#8220;in silico&#8221; experiments, ethical restrictions may not apply, which has led <a href="https://proceedings.mlr.press/v202/aher23a/aher23a.pdf">Aher et al. (2023)</a> to repeat <a href="https://psycnet.apa.org/record/1964-03472-001">Milgram&#8217;s (1963)</a> controversial shock experiment,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> finding similar results.</p><p>Borth Argyle et al. (2022) and Aher et al. (2023) capture the excitement among (some) social scientists about how AI may superpower their research. This excitement is likely accentuated by challenges faced in the social sciences. For example, survey-based data collection is in <a href="https://www.youtube.com/watch?v=h0ffIJ7ZO4U">dire straits</a>. From 1997 to 2012, Pew&#8217;s research household-level response rate dropped from 37% to 12% <a href="https://law.osu.edu/electionlaw/litigation/documents/Veasey7885.pdf">(Kohut et al., 2012)</a>. Further, concerns about the generalizability and utility of modern social science have led to numerous calls for reform [e.g., see <a href="https://www.nature.com/articles/s41562-016-0015">Watts (2017)</a> and <a href="https://www.cambridge.org/core/journals/behavioral-and-brain-sciences/article/beyond-playing-20-questions-with-nature-integrative-experiment-design-in-the-social-and-behavioral-sciences/7E0D34D5AE2EFB9C0902414C23E0C292">Almaatouq (2022)</a>]. Could AI be the panacea for all these problems?</p><p>A lot of people are hesitant to call it a day. The extent to which and the reasons why being very diverse. To some, <a href="https://kevinmunger.substack.com/p/i-strongly-feel-that-this-is-an-insult">the idea is insulting at a fundamental level</a>. Social scientists study humans, so why should they, in all their intricacies, be replaced? I perceive this refusal to be akin to how <a href="https://www.nytimes.com/2022/09/02/technology/ai-artificial-intelligence-artists.html">many artists loathe AI-generated art</a>; the overall feeling is best described by <a href="https://www.youtube.com/watch?v=ngZ0K3lWKRc">this video</a> of Miyazaki&#8217;s response to an AI demo being shown to him. He says</p><blockquote><p><em>I would never wish to incorporate this technology into my work at all. I strongly feel that this is an insult to life itself.</em></p></blockquote><p>However, concerns around AI-powered research go far beyond a visceral feeling that &#8220;something is off.&#8221; The last &#8220;scientific revolution&#8221; triggered by AI, researchers using traditional machine learning algorithms, led to a lot of bad science [e.g., many of the results are incorrect due to data leakage<a href="https://www.cell.com/patterns/fulltext/S2666-3899(23)00159-9?uuid=uuid%3Aa1520953-3278-4f02-baaa-71332a2e9ed6"> (Sayash and Narayanan, 2023)</a>] &#8212; and so could the new wave of generative AI. Without the hurdles of recruiting and meaningfully engaging with human subjects, &#8220;silicon samples&#8221; could lead to a flurry of lazy studies. A significant problem with modern science is that increasing the production of scientific artifacts (e.g., papers, code, grants) does not necessarily increase our understanding of the world &#8212; and the likes of ChatGPT may make this much worse.</p><p>But beyond the explosion of lazy studies, AI could hinder the production of scientific knowledge in other, more subtle ways.<a href="https://www.nature.com/articles/s41586-024-07146-0"> Messeri and Crocket (2024)</a> argue that the widespread use of AI might hinder science by creating the illusion that we understand more about the world than we do. In their words:</p><blockquote><p><em>AI solutions can also exploit our cognitive limitations, making us vulnerable to illusions of understanding in which we believe we understand more about the world than we actually do. Such illusions obscure the scientific community&#8217;s ability to see the formation of scientific monocultures, in which some types of methods, questions and viewpoints come to dominate alternative approaches, making science less innovative and more vulnerable to errors.</em></p></blockquote><p>Last, and perhaps most important, it turns out that AI is far from perfect at simulating human behavior. After late-breaking results from Argyle et al. (2022), many papers found that LLMs are actually quite bad at simulating the diversity of human behavior.</p><p>A particularly meaningful mode of failure is what <a href="https://aclanthology.org/2023.emnlp-main.669.pdf">Cheng et al. (2023)</a> refer to as <em>caricatures: </em>&#8220;an exaggerated narrative of the persona (the demographic that we aim to simulate) rather than a meaningful response to the topic.&#8221; Humans are nuanced, complex, and sometimes just plainly inconsistent &#8212; and LLM simulations are not capturing those &#8220;wrinkles.&#8221; <a href="https://arxiv.org/pdf/2402.01908#page=9.82">Wang et al. (2024)</a> show firsthand that simple persona prompting fails to capture group heterogeneity. Their assessment is quite dire and well reflected in their title:<em> Large language models should not replace human participants because they can misportray and flatten identity groups</em>.</p><p>But to what extent can we make claims about LLMs&nbsp;<em>in general</em>&nbsp;when we are prompting them in the simplest ways possible? <a href="https://aclanthology.org/2024.acl-long.554.pdf">Hu and Colier (2024)</a> find that, actually, increasing the number of persona variables included in the prompt makes predictions much more accurate. Further, the more variables are correlated with the outcome we want to predict, the better:</p><blockquote><p><em>We find a linear relationship in our setting: the more persona variables are correlated with the outcome variable, the better LLMs predictions are using persona prompting. Large, preference-tuned models perform best and can explain up to 81% of variance found in human responses. However, when the utility of persona variables is low, persona prompting has little effect.</em></p></blockquote><p>So, for example, maybe saying that someone is <em>&#8220;a 53-year-old Hispanic-American woman who identifies as a Democrat&#8221; </em>is not enough! Perhaps we need variables and data that capture nuances and the complexities of human subjects. But fortunately, LLMs can extract these nuances from unstructured text. In groundbreaking work, <a href="https://arxiv.org/pdf/2411.10109?">Park et al. (2024)</a> show that LLMs can consistently simulate people when prompted not with demographic variables but with long-form, unstructured interviews. They write:</p><blockquote><p><em>To create simulations that better reflect the myriad, often idiosyncratic, factors that influence individuals' attitudes, beliefs, and behaviors, we turn to in-depth interviews&#8212;a method that previous work on predicting human life outcomes has employed to capture insights beyond what can be obtained through traditional surveys and demographic instruments.</em></p></blockquote><p>I believe this approach captures what&#8217;s most exciting about simulating human behavior with AI. These simulations open new opportunities to <em>augment</em> rather than replace existing practices in the social sciences. Even though AI is not a panacea for studying human behavior, this early work indicates that it will be a helpful building block for creating solutions to existing challenges!</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>I am deliberately avoiding the &#8220;what is intelligence?&#8221; debate here. Simulating something that looks human is interesting by itself (and useful enough).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>The description of Aher et al. (2023) here is quite good <em>&#8220;An authority figure in the subject&#8217;s eyes, orders the subject to shock a victim with increasingly high voltage shocks. After receiving 20 shocks, the victim (an actor in another room) starts banging on the wall and refuses to participate but the experimenter urges the subject to keep shocking the victim. Milgram found that many subjects completed administering 30 shocks, showing a surprisingly strong level of compliance for following the malevolent instructions of an authority figure who had no special powers to enforce his commands.&#8221;</em></p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Why study "Machine Behavior"?]]></title><description><![CDATA[Or "Why Machine Behavior is a Nice Buzzword"]]></description><link>https://doomscrollingbabel.manoel.xyz/p/why-study-machine-behavior</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/why-study-machine-behavior</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Sat, 28 Dec 2024 11:43:22 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/323326e0-bb41-4411-a7ce-809d05f9e963_1014x670.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Artificial Intelligence (AI) algorithms have become <strong>ubiquitous</strong>. From dating apps to autonomous weapons, the boundaries of what is and what isn&#8217;t AI blurs as applications using deep neural networks and decision trees become commonplace<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> and widely used across the sciences <a href="https://www.nature.com/articles/s41562-024-02020-5">(Gao and Wang, 2024)</a>. With the rise of Generative AI, these texts, images, and videos are increasingly AI-generated. This has even changed how we speak, with words distinctively associated with ChatGPT increasing in presentations, talks, and speeches <a href="https://arxiv.org/abs/2409.01754">(Yakura et al., 2024)</a>.</p><p>As AI becomes ubiquitous, AI technologies have also become <strong>consequential </strong>to society. Standalone AI algorithms can be consequential by themselves; slight tweaks to criminal risk prediction algorithms have a tangible impact on the lives of incarcerated people [as these algorithms are used in court; see <a href="https://www.propublica.org/article/how-we-analyzed-the-compas-recidivism-algorithm">Larson et al. (2016)</a>]. However, the effects of AI become even more remarkable if we consider what it enables and augments.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><ul><li><p><strong>Some things can only exist due to advancements in AI:</strong> TikTok does not exist without recommender systems; it is not that AI &#8220;changed&#8221; TikTok; the whole idea behind it fundamentally relies on a recommender system. Without recommender systems, <em>TikTok isn&#8217;t</em>.</p></li><li><p><strong>Other things changed drastically due to advancements in AI:</strong> Content moderation in social media platforms predates AI. <a href="https://en.wikipedia.org/wiki/Slashdot">Slashdot</a> (1997) had content moderation. However, AI completely changed how content moderation is done in modern social media platforms like Facebook, Reddit, and Twitter, with a large swathe of content removals automatically triggered by internal classifiers <a href="https://dl.acm.org/doi/abs/10.1145/3543507.3583275?casa_token=greQIxvbNXcAAAAA:ivBWipTZ0KaDIlg3S0d_RtRLiM-vp_yY_UH4Ih-Q5-f7ibYHtsrCswrSthPgWax6TXAWKrfo0xI">(Horta Ribeiro et al., 2023</a>).</p></li></ul><p>In light of the ubiquitousness and consequentially of AI, scholars have called for a research program that studies the behavior of AI (or of &#8220;machines&#8221;) <em>within</em> human society.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>As <a href="https://www.nature.com/articles/s41586-019-1138-y">Rahwan et al. (2019)</a> put it:</p><blockquote><p><em>In his landmark 1969 book Sciences of the Artificial, Nobel Laureate Herbert Simon wrote: &#8220;Natural science is knowledge about natural objects and phenomena. We ask whether there cannot also be &#8216;artificial&#8217; science&#8212;knowledge about artificial objects and phenomena.&#8221; In line with Simon&#8217;s vision, we describe the emergence of an interdisciplinary field of scientific study. This field is concerned with the scientific study of intelligent machines, not as engineering artefacts, but as a class of actors with particular behavioural patterns and ecology. This field overlaps with, but is distinct from, computer science and robotics. It treats machine behaviour empirically. This is akin to how ethology and behavioural ecology study animal behaviour by integrating physiology and biochemistry&#8212;intrinsic properties&#8212;with the study of ecology and evolution&#8212;properties shaped by the environment. Animal and human behaviours cannot be fully understood without the study of the contexts in which behaviours occur. Machine behaviour similarly cannot be fully understood without the integrated study of algorithms and the social environments in which algorithms operate.</em></p></blockquote><p>But it is worth asking: why do we need this new &#8220;field&#8221;? What are the shortcomings of previous research that can&#8217;t explain the brand-new world of recommender systems (and LLMs)?</p><p><a href="https://www.nature.com/articles/s41586-021-03666-1">Wagner et al. (2021)</a> provide a partial answer: understanding all societal phenomena is incredibly challenging when algorithms are co-shaping society. They identify a couple of challenges with doing social science in what they call &#8220;algorithmically infused societies.&#8221; Namely, when algorithms enter the playground: #1) social science theories crumble; #2) measuring things becomes harder; #3) (mis)measuring things have consequences.</p><p>I illustrate the above three with an example from my own line of research: trying to study the impact of recommender systems on social media. As you might already have heard, there is widespread concern that recommender systems in social media platforms like YouTube or Facebook would radicalize people <a href="https://coinse.github.io/assets/files/teaching/cs489/Tufekci.pdf">(Tufekci, 2018)</a>. And, indeed, a lot of people were radicalized &#8220;within&#8221; social media&#8212;that is undeniable.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> But the question remains: <em>what is the role of the algorithm? </em>Let&#8217;s see if points #1, #2, and #3 by Wagner et al. (2021) hold up here.</p><p><strong>Point #1. </strong>Radicalization is well-studied within the social sciences. However, radicalization in the age of social media is different, as individuals creating &#8220;extreme&#8221; content had to engage with algorithmic feeds and algorithmic content moderation. This has led to a lot of confusion around the impact of recommender systems. The conclusion that eventually emerged after many years of debate and studies is nicely summarized by <a href="https://journals.sagepub.com/doi/abs/10.1177/1940161220964767">Munger and Phillips (2022)</a>:</p><p><em>(...) we argue for the need to study the YouTube Right systematically an advance a &#8220;supply-and-demand framework&#8221; to understand the proliferation of rightwing media on the platform. To date, journalistic and scholarly work has argued that YouTube&#8217;s recommendation algorithm has led viewers to extremist content, radicalizing them to further-right views. We believe that this conclusion is premature, and we are certain that this is not the only important research question to be asked by political scientists about right-wing content on YouTube, or YouTube more broadly.</em></p><p><strong>Point #2. </strong>This confusion was primarily caused by how hard it is to measure &#8220;online radicalization&#8221; to begin with. For example, when studying radicalization, most studies [including Horta Ribeiro et al. (2020)] did the most na&#239;ve thing possible. They measured the extent to which extreme content was recommended as you randomly walked through the (dynamic) recommendation graph. However, these measurements assume no &#8220;user models&#8221;. They assume no meaningful interaction pattern between humans and machines. In subsequent work, when human interaction patterns are considered, there&#8217;s little evidence of any &#8220;algorithmic radicalization&#8221; taking place <a href="https://www.pnas.org/doi/10.1073/pnas.2313377121">(Hosseinmardi et al. 2024)</a>.</p><p><strong>Point #3. </strong>Mismeasurements have consequences! In this case, it led to the deprioritization of important content moderation and platform governance policies. The mismeasurement of algorithmic radicalization arguably steered people away from other, more important interventions. However, it is worth mentioning that Wagner et al.&#8217;s point here is even broader. Even <em>correct</em> measurements have consequences, as they can feed back into algorithms that shape human behavior. In our running example, there are huge efforts from online companies <em>not to recommend</em> extreme content. Therefore, this whole literature around radicalization happened while YouTube was doing its best (I guess) to recommend as little extreme content as possible. It also coincided with a tightening of content moderation practices within these platforms. Many original channels studied by Horta Ribeiro et al. (2020) were banned shortly after the study was published and got much media attention (although a causal relationship was never established).</p><div><hr></div><p>Both Rahwan et al. (2019) and Wagner et al. (2021) were written before the AI explosion in late 2022 with the popularization of ChatGPT. Ever since, there has been increasing interest in AI alignment, in using LLMs to simulate human behavior, and even in the consequences of the deployment of AI agents into the wild.</p><p>I&#8217;d argue that machine behavior is a nice buzzword to refer to the study of these phenomena. On the one hand, if we see AI as merely an engineering problem, we risk overlooking the empirical reality that these systems shape&#8212;and are shaped by&#8212;the very social environments in which they operate. On the other hand, if we ignore the role of AI, we risk overlooking the technical underpinnings and design choices that critically shape how users behave and interact in digital environments. Only by acknowledging the messy interplay between humans and machines can we fully understand&#8212;and responsibly guide&#8212;the future of AI in society.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Tesler&#8217;s theorem even states that &#8220;AI is whatever hasn't been done yet;&#8221; see <a href="https://en.wikipedia.org/wiki/AI_effect">AI effect</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Another cool paper here is <a href="https://www.nature.com/articles/s41562-024-02001-8?fromPaywallRec=false">Tsvetkova et al. (2024)</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p><a href="https://dl.acm.org/doi/10.1145/3351095.3372879">Horta Ribeiro et al. (2020)</a>, yours truly, show how users consistently migrated from contrarian to extreme content in the late 2010s. This is not to say that the algorithms were to blame, although this was way less clear in 2019 when this was written.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Is Facebook “Standard” Algorithm Polarizing?]]></title><description><![CDATA[Were the findings of Meta's Science paper on algorithmic effects debunked?]]></description><link>https://doomscrollingbabel.manoel.xyz/p/is-facebook-standard-algorithm-polarizing</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/is-facebook-standard-algorithm-polarizing</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Mon, 30 Sep 2024 11:37:09 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/ecd8785b-a7b6-4a69-afac-de38f602086f_1024x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Guess et al. (2023) showed in <a href="https://www.science.org/doi/10.1126/science.abp9364">a prominent paper in Science</a> that switching some users to a chronological (rather than algorithmic) feed for three months during the election in 2023 &#8220;did not significantly alter levels of issue polarization, affective polarization, political knowledge, or other key attitudes.&#8221; However, in <a href="https://www.science.org/doi/10.1126/science.abp9364#elettersSection">a new correspondence</a>, also in Science,&nbsp;Bagchi et al. (2024) question the validity of the prominent study. Meta changed the algorithm before the elections, which would call into question the study&#8217;s findings. Most specifically, Bagchi et al. (2024) argue the study may mislead readers&nbsp;&#8220;to conclude that the Facebook news feed algorithm used outside of the study period mitigates political misinformation compared to (the) reverse chronological feed.&#8221;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!frF6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F538935d6-f85a-4da9-b27b-381879ee8558_1430x734.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!frF6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F538935d6-f85a-4da9-b27b-381879ee8558_1430x734.png 424w, https://substackcdn.com/image/fetch/$s_!frF6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F538935d6-f85a-4da9-b27b-381879ee8558_1430x734.png 848w, https://substackcdn.com/image/fetch/$s_!frF6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F538935d6-f85a-4da9-b27b-381879ee8558_1430x734.png 1272w, https://substackcdn.com/image/fetch/$s_!frF6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F538935d6-f85a-4da9-b27b-381879ee8558_1430x734.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!frF6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F538935d6-f85a-4da9-b27b-381879ee8558_1430x734.png" width="1430" height="734" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/538935d6-f85a-4da9-b27b-381879ee8558_1430x734.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:734,&quot;width&quot;:1430,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!frF6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F538935d6-f85a-4da9-b27b-381879ee8558_1430x734.png 424w, https://substackcdn.com/image/fetch/$s_!frF6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F538935d6-f85a-4da9-b27b-381879ee8558_1430x734.png 848w, https://substackcdn.com/image/fetch/$s_!frF6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F538935d6-f85a-4da9-b27b-381879ee8558_1430x734.png 1272w, https://substackcdn.com/image/fetch/$s_!frF6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F538935d6-f85a-4da9-b27b-381879ee8558_1430x734.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Bagchi et al. (2024) base their claims on an analysis of a <a href="https://solomonmg.github.io/pdf/Facebook_DP_URLs_Dataset.pdf">large dataset released by Meta</a> containing the amount of time users viewed, clicked, and shared URLs on Facebook. Using a list of outlets known for low-quality reporting (<a href="http://media bias fact check">Media Bias Fact Check</a>), they find a drop starting in early November 2020 that goes until early March 2021. These coincide with changes to the Facebook algorithm in November. Without such changes, Guess et al. (2023) may have found a positive effect!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Authors responded with their own letter, hereinafter Guess et al. (2024). They make three key arguments against Bagchi et al. (2024), all of which I found reasonable:</p><ul><li><p>First, the experiment's internal validity is not impacted by these changes. Accurate causal conclusions can be reached for the specific period analyzed. Facebook did enact these changes, after all. </p></li><li><p>Second, the data analyzed by Bagchi et al. (2024) is incomplete, it only contains URLs shared more than 100 times, but not actual content posted on Facebook (which could be actually misleading). When analyzing Guess et al. (2023) data, they found little change in the number of unreliable sources pre- vs. post-study period. In other words, in their control group (where the recommender algorithm is enabled), they don&#8217;t observe this drop in the fraction of untrustworthy content. </p></li><li><p>Third, that the evidence used by Bagchi et al. (2024) is not causal. The &#8220;information ecosystem&#8221; was absolutely crazy between November 2020 and March 2021. Observed changes might have come from the many things that were going on, e.g., Biden got elected.&nbsp;</p></li></ul><h2>Things I wish the letters had done</h2><p>After reading each letter, I wished that more analyzes had been conducted.&nbsp;</p><p>When reading Bagchi et al. (2024), I found in their references <a href="https://www.washingtonpost.com/documents/5bfed332-d350-47c0-8562-0137a4435c68.pdf#page=35">a fairly precise description of algorithmic changes enacted by Facebook</a>. This was obtained by the January 6th Committee (but left out of the final report). Most important:</p><blockquote><p><em><strong>(A)</strong> Filter low News Ecosystem Quality (NEQ) pages from Pages you May Like to prevent low quality and misinformation pages from becoming viral.</em></p><ul><li><p><em>Launched 10/22</em></p></li><li><p><em>Reduced to 75% on 12/1, then 50% on 12/3, 25% on 12/8, and deprecated on 12/10.</em></p></li><li><p><em>Relaunched in response to Jan 6th.</em></p></li></ul><p><em><strong>(B) </strong>Deploy the virality circuit breaker, which prevents the likelihood of URLs from new or unknown external domains that may contain misinformation from being boosted</em></p><ul><li><p><em>10/9 launched at 100x threshold</em></p></li><li><p><em>10/23 launched at 25x threshold</em></p></li><li><p><em>12/1 reduced to 75%, then 50% on 12/3, then 25% on 12/8</em></p></li><li><p><em>Deprecated on 12/10</em></p></li></ul><p><em><strong>(C)</strong> Demote content from users who posted multiple pieces of third-party fact-checked misinformation in the past 30 days.</em></p><ul><li><p><em>Launched 11/5</em></p></li><li><p><em>Reduced to 50% on 12/2, then deprecated 12/3</em></p></li><li><p><em>Relaunched 1/14</em></p></li><li><p><em>Deprecated 1/29.</em></p></li></ul><p><em><strong>(D)</strong> Demote low NEQ news and boost high NEQ news in order to increase the average quality of news in connected news feed</em></p><ul><li><p><em>Launched 11/7</em></p></li><li><p><em>Reduced to 75% on 12/1, then 50% on 12/3, 25% on 12/8, and deprecated 12/10</em></p></li><li><p><em>Relaunched 1/13</em></p></li><li><p><em>Deprecated 2/16</em></p></li></ul></blockquote><p>So, basically, two of the relevant changes (A and B) above were deployed in October. This does not coincide with the sudden increase in trustworthy news in early November observed by Bagchi et al. (2024). Two relevant changes (B and D) were deployed in early November but were deprecated in December! If they were driving the consumption of trustworthy news, I would expect that the consumption of trustworthy news would have dropped again in December.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!K6l5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa654c317-268b-4a26-b8a6-e3d42fafc9c5_976x1080.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!K6l5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa654c317-268b-4a26-b8a6-e3d42fafc9c5_976x1080.png 424w, https://substackcdn.com/image/fetch/$s_!K6l5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa654c317-268b-4a26-b8a6-e3d42fafc9c5_976x1080.png 848w, https://substackcdn.com/image/fetch/$s_!K6l5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa654c317-268b-4a26-b8a6-e3d42fafc9c5_976x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!K6l5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa654c317-268b-4a26-b8a6-e3d42fafc9c5_976x1080.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!K6l5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa654c317-268b-4a26-b8a6-e3d42fafc9c5_976x1080.png" width="976" height="1080" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a654c317-268b-4a26-b8a6-e3d42fafc9c5_976x1080.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1080,&quot;width&quot;:976,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!K6l5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa654c317-268b-4a26-b8a6-e3d42fafc9c5_976x1080.png 424w, https://substackcdn.com/image/fetch/$s_!K6l5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa654c317-268b-4a26-b8a6-e3d42fafc9c5_976x1080.png 848w, https://substackcdn.com/image/fetch/$s_!K6l5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa654c317-268b-4a26-b8a6-e3d42fafc9c5_976x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!K6l5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa654c317-268b-4a26-b8a6-e3d42fafc9c5_976x1080.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>When reading Guess et al. (2024), I wondered why they wouldn&#8217;t simply rerun the analysis in Figure 1 (reproduced above), considering the (short) period <em>before</em> the changes to the algorithm were enacted. In other words, couldn&#8217;t they have redone the analyses considering the month <em>after</em> the start of the experiment (in late September) and <em>before</em>&nbsp;the changes pointed out by Bagchi et al. (2024)? This would have been a simple, elegant way to determine how much the algorithmic changes mattered.</p><h2>But what is external validity here, anyway?</h2><p>Guess et al. (2023) used pretty timid language to describe the findings, indicating that it was hard to generalize them. However, the lack of external validity is inherent to the problem of algorithmic effects! Let me explain. Implicit in both the letter by Bagchi et al. (2024) and in an editorial discussing the letter exchange is this idea of a &#8220;canonical&#8221; Facebook algorithm (the editorial calls it the &#8220;site&#8217;s default algorithm&#8221;). But Facebook has no default algorithm; the algorithm is constantly changing due to new data being posted on the website and new tweaks being made to the (dozens) of models used.&nbsp;</p><p>Bagchi et al. (2024) are correct in two senses. First, the original study by Guess et al. (2023) original study would have been even stronger if they studied the pre-changes algorithm. Second, the original should have mentioned the algorithmic changes made for the election. Their study would have been stronger if they had done the kinds of sanity checks they did in their letter, to begin with (and even stronger if they redid the analysis considering periods with different algorithmic &#8220;regimes&#8221;). However, they are misleading in calling these changes a threat to the study's overall validity. If anything, they are an incremental threat to its external validity, which was already questionable (by the authors themselves).</p><h2>So what?</h2><p>Outside the ivory tower, however, things get a bit weird. Meta&#8217;s Nick Clegg (President of Global Affairs) said that the original paper undermined claims that the site was designed to &#8220;serve people content that keeps them divided,&#8221; which is clearly false. The social dynamics induced by platforms like Facebook transcend their algorithm. In that context, I guess that the letter may serve as a reminder that more work is needed to study the effect of social media (and of algorithms) on society.</p><p>But we must be careful! This letter and the reaction to it may also lead to a misleading narrative. The <a href="https://www.ucd.ie/newsandopinion/news/2024/september/27/neweletterinsciencedebunksmeta-fundedstudysuggestingitsnews-feedalgorithmsarenotmajordriversofmisinformation/">press release</a> from University College Dublin had the sensationalist title &#8220;New eLetter in Science debunks Meta-funded study suggesting its news-feed algorithms are not major drivers of misinformation.&#8221; It also contains a quote by one of the authors of the letter that is absolutely not supported by the letters: &#8220;Our results show that social media companies can mitigate the spread of misinformation by modifying their algorithms but may not have financial incentives to do so.&#8221; This is false! The letter adds incremental concerns about the external validity of the study. That&#8217;s it.</p><p>One reasonable question to ask here is: what does the study (and the letter) tell us that will be helpful for the future? I share <a href="https://kevinmunger.substack.com/p/is-the-best-social-science-good-enough">Kevin Munger&#8217;s take</a> that we should think about these things in a Bayesian sense. The original Science study indicates that the Facebook algorithm will not be a key driver of polarization in the next election. The letter, indicates that, in absence of special changes, we might expect it to have a slightly bigger effect. In my view, the evidence of the Science study is stronger than that of the letter, given the response by Guess et al. 2024.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2><strong>Additional comments</strong></h2><p> <a href="https://substack.com/profile/129436715-przemyslaw-grabowicz">Przemyslaw Grabowicz</a>, the corresponding author of the letter, wrote an interesting comment on this post. They are in the comments just below, but I append them here:</p><blockquote><p>Manoel, I greatly appreciate your feedback, since it makes me realize that I should write more about our Science eLetter and its implications. Once I find time, I'll put together a proper overall response at <a href="http://uncommongood.substack.com/">UncommonGood.substack.com</a>. Meanwhile, let me quickly respond to the "three key arguments against Bagchi et al. (2024)" that you've taken from the response eLetter by Guess et al.</p><p>First, I don't think the paper by Guess et al. is internally valid. Consider the following. If a paper computes a causal estimate based on an experiment, but the control condition is meaningfully changed during the experiment specifically to affect the target causal estimand, and the paper doesn't reveal anything about that specific change, nor accounts for it, then the change could result in any desirable value of the causal estimate, without revealing anything about how it happened. In other words, a causal claim must define exactly what control and treatment conditions are. If it doesn't then the causal claim may be invalid in the situations where the description misses something meaningful that's related to the estimand. If the change was not described, then the default assumption should be that there was no meaningful change during the experiment. However, during the the experiment of Guess et al. there was a meaningful change introduced...</p><p>Second, you write that:</p><p>&gt; Guess et al. (2023) data, they found little change in the number of unreliable sources pre- vs. post-study period. In other words, in their control group (where the recommender algorithm is enabled), they don&#8217;t observe this drop in the fraction of untrustworthy content.</p><p>Ok, so let's see what exactly Guess et al. write in their response eLetter, I quote</p><p>&gt; Over the 90 days prior to the treatment, untrustworthy sources represented 2.9% of all content seen by participants in the Algorithmic Feed (control) group &#8211; during the study period, this dropped only modestly to 2.6%.</p><p>So, according to their own measure, there was a drop in the fraction of misinformation from 2.9% to 2.6%, so that's 10.5% relative drop (0.3/2.9), whereas we reported a 24% drop. Note, however, that only about half of their treatment period overlaps with the period of Facebook's emergency interventions. If it overlapped entirely, then probably instead of 10.5% drop, we would observe a 21% drop. That's starts to be quite close to the 24% drop we measured using a different dataset and a different notion of misinformation.</p><p>Third, you're right that the evidence used by Bagchi et al. (2024) is not causal. However, in our eLetter we haven't made any causal statements. Instead, we're pointing out that Guess et al. made causal statements without properly describing the control condition of its experiment. That said, we provided also potential explanations for the drop in the fraction of misinformation in news feeds of users. This explanation aligns with the reasons why emergency measures were introduced. These reasons were provided both officially by Facebook representatives [1], and unofficially by Facebook employees and a whistleblower, Francis Haugen [2, 3].</p><p>[1] <a href="https://www.nytimes.com/2020/12/16/technology/facebook-reverses-postelection-algorithm-changes-that-boosted-news-from-authoritative-sources.html">https://www.nytimes.com/2020/12/16/technology/facebook-reverses-postelection-algorithm-changes-that-boosted-news-from-authoritative-sources.html</a></p><p>[2] <a href="https://www.wsj.com/articles/the-facebook-files-11631713039">https://www.wsj.com/articles/the-facebook-files-11631713039</a></p><p>[3] <a href="https://www.washingtonpost.com/documents/5bfed332-d350-47c0-8562-0137a4435c68.pdf">https://www.washingtonpost.com/documents/5bfed332-d350-47c0-8562-0137a4435c68.pdf</a></p></blockquote><p>To which I replied:</p><blockquote><p>Thanks for the reply, I attached it to the end of the post!</p><p>I disagree with point #1: the control group was "Facebook as it was during the election," that's fine. It is like saying that you an experiment to study in mobility in a city is invalid because there were changes due to Christmas &#8212; most likely than not, Facebook will always have changes for US elections... </p><p>I am not super convinced of point #2 on either way. You make a good point about the "entire treatment period". But still, this is such a convoluted thing because there are exogenous shocks to the demand and supply of news here. </p><p>I agree with you on point #3. You folks didn't do any causal claims in the letter, but note that: &#8220;Our results show that social media companies can mitigate the spread of misinformation by modifying their algorithms but may not have financial incentives to do so&#8221; is a causal statement, which is what bothered me.</p></blockquote><p></p>]]></content:encoded></item><item><title><![CDATA[Discourse on social media is vibes-based]]></title><description><![CDATA[In the fall of 2021, The Wall Street Journal dropped a bombshell on Facebook (now Meta): &#8220;Facebook Knows Instagram Is Toxic for Teen Girls.&#8221; The report drew from &#8220;The Facebook Files,&#8221; documents leaked by former employee Frances Haugen, which, among other things, discussed internal research on mental health and well-being.]]></description><link>https://doomscrollingbabel.manoel.xyz/p/discourse-on-social-media-is-vibes</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/discourse-on-social-media-is-vibes</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Tue, 13 Aug 2024 14:45:53 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/1e12e395-f0f0-48e2-b805-21144b60bb7b_1024x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In the fall of 2021, The Wall Street Journal dropped a bombshell on Facebook (now Meta): <em>&#8220;<a href="https://www.wsj.com/articles/facebook-knows-instagram-is-toxic-for-teen-girls-company-documents-show-11631620739">Facebook Knows Instagram Is Toxic for Teen Girls</a></em><a href="https://www.wsj.com/articles/facebook-knows-instagram-is-toxic-for-teen-girls-company-documents-show-11631620739">.</a>&#8221; The report drew from &#8220;The Facebook Files,&#8221; documents leaked by former employee Frances Haugen, which, among other things, discussed internal research on mental health and well-being. One in five teenagers, the documents suggested, said Instagram made them feel worse about themselves (in a UK/US sample of around a thousand users).</p><p>Fast forward to the spring of 2024, Jonathan Haidt&#8217;s book &#8220;The Anxious Generation&#8221; made it to the top charts, sparking widespread conversation about the mental health crisis among young people. Haidt&#8217;s argument very much resonated with the internal research leaked by Haugen: there is a strong correlation between the increase in social media use and the worsening of mental health outcomes among teenagers. His conclusion: the brave new world of social media and mobile phones is harming teens&#8217; mental health.</p><p>Told in this fashion, one could believe that the reaction to these two notable events would be similar; after all, there&#8217;s so much in common between the arguments put forth by Haidt and by Meta internal researchers. They both point to the suggested negative impact of social media on teen&#8217;s mental health, but more so, they both fall short of cleanly identifying a causal effect.&nbsp;</p><p>In the case of Facebook's internal documents, data was entirely self-reported, which might have led participants to list Instagram as the culprit for their mental health issues simply because this notion is already widespread in pop culture (but not in academia). In Haidt&#8217;s book, conclusions are drawn from aggregated trends of tremendously complex outcomes, making it particularly hard to discern correlation from causation. All in all, in both cases, it may very well be that &#8220;<em>misery causes Facebook</em>&#8221; (this pearl is by Kevin Munger), and the studies are off.</p><p>However, the point is not to dunk on Facebook's internal documents or Haidt&#8217;s book&#8212;studying the causal effect of social media on well-being is remarkably challenging, and it is not as if they overlooked an obvious research design or method that would improve the credibility of their results. Rather, my point is to analyze the reaction to these two events. The conclusions drawn from &#8220;The Facebook Files&#8221; largely agree with the &#8220;Facebook Bad&#8221; trope created after the Cambridge Analytica scandal and drew little backlash for their limitations. On the other hand, Haidt&#8217;s book was met with extreme skepticism; E.g., a <a href="https://www.nature.com/articles/d41586-024-00902-2">book review</a> in Nature by a leading psychologist in the area compares Haidt&#8217;s arguments to drawing similar lines on the first day of a statistics course.</p><p>But why were the reactions so different? I hypothesize that it is all about <em>The Vibes</em>, a term coined by no one other than the anxious generation. When The Wall Street Journal published its report, it did so against mounting skepticism and distrust toward Big Tech, particularly Facebook. In contrast, when assessing Haidt&#8217;s book, critics could not detach his argument from his position as a critic of progressive ideologies. Looking at no trend lines, I dare to say that what caused the &#8220;differential criticism&#8221; was simply Haidt&#8217;s vibes.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Content Curation in Online Platforms]]></title><description><![CDATA[(This is a big rant on why research on content moderation, algorithms, and monetization strategies is hard and why we desperately need it.]]></description><link>https://doomscrollingbabel.manoel.xyz/p/content-curation-in-online-platforms</link><guid isPermaLink="false">https://doomscrollingbabel.manoel.xyz/p/content-curation-in-online-platforms</guid><dc:creator><![CDATA[Manoel Horta Ribeiro]]></dc:creator><pubDate>Sat, 06 Apr 2024 18:03:41 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/8fbb97e0-4b34-4a61-bbf1-c5a31f835dba_1500x1008.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://doomscrollingbabel.manoel.xyz/subscribe?"><span>Subscribe now</span></a></p><p><em>(This is a big rant on why research on content moderation, algorithms, and monetization strategies is hard and why we desperately need it. It is an interpolation between some of the materials I prepared for my job talk and my PhD thesis)</em></p><div><hr></div><p>Online platforms like Facebook, Wikipedia, YouTube, Amazon, Uber, DoorDash, Airbnb, and Tinder have changed the world and become embroidered into the social fabric. It is hard to imagine how our lives would be without them: our economies, our relationships, and how we acquire knowledge have become deeply connected to these online platforms. The United Nations Conference on Trade and Development estimated that the global value of e-commerce sales reached almost <a href="https://unctad.org/publication/covid-19-and-e-commerce-global-review">26 trillion dollars in 2018</a>. Pew Research estimated that <a href="https://www.pewresearch.org/short-reads/2023/02/02/key-findings-about-online-dating-in-the-u-s/">around 10% of partnered US adults met their match online dating in 2023</a>. Wikipedia received over <a href="https://go.epfl.ch/wikipedia_stats">7 billion monthly visits</a> in 2023, satisfying users with the most diverse information needs. </p><p>Thus, it is perhaps not surprising that online platforms are also strongly connected to some of the most significant societal challenges of the 21st century. E-commerce platforms are responsible for <a href="https://pubs.acs.org/doi/10.1021/acs.est.2c00299">a sizeable chunk of greenhouse gas emissions</a>. Gig work platforms like Uber and DoorDash ignited discussions about workers' rights and precarious employment. Radicalization and terrorism have become an online-first phenomenon: mainstream social media platforms like YouTube saw an influx of radical content that snowballed in popularity in the late 2010s, and fringe platforms like Gab and Parler were tightly associated with terrorist attacks and anti-democratic protests.</p><p>However, online platforms are not &#8220;immovable rocks&#8221; that we should accept as they are; they are sociotechnical systems where design choices, policies, and algorithms steer human behavior. And in the spirit of <a href="https://www.sfu.ca/~palys/Campbell-1991-MethodsForTheExperimentingSociety.pdf">Campbell's experimenting society</a>, we should propose and assess ways of improving online platforms, maximizing their benefits, and minimizing their harms. </p><p>The critical enterprise of online platforms is that they <em>curate content</em>. Users on these platforms upload images, list products to sell, and make profiles. At the same time, platforms curate this content and serve it to other users on the platform. And there are some critical ways in which they do so: they decide what to recommend to users, how to monetize content on the platform, and what is allowed on the platform.</p><p>Content curation practices have captured the imagination of journalists, politicians, and the general public. For example, tech CEOs testified in Senate Hearings in <a href="https://www.washingtonpost.com/news/the-switch/wp/2018/04/10/transcript-of-mark-zuckerbergs-senate-hearing/">2018</a>, <a href="https://www.npr.org/2020/10/28/928532702/facebook-twitter-google-ceos-testify-to-senate-what-to-watch-for">2020</a>, <a href="https://www.npr.org/2021/03/25/980510388/facebook-twitter-google-ceos-testify-before-congress-4-things-to-know">2021</a>, and <a href="https://apnews.com/article/meta-tiktok-snap-discord-zuckerberg-testify-senate-00754a6bea92aaad62585ed55f219932">2024</a>; they were often asked to discuss content curation practices like recommender systems or content moderation. In a 2018 poll, <a href="https://www.cnbc.com/2018/09/06/twitter-permanently-bans-alex-jones-and-infowars-accounts.html">65% of self-described US conservatives thought social media platforms were censoring conservative ideas</a>. In that context, research informing content curation practices is <em>actionable</em>; it can propose concrete ways of changing online platforms and inform stakeholders. For example, deplatforming or banning individuals or collectives from our online ecosystem has <a href="http://theconversation.com/banning-disruptive-online-groups-is-a-game-of-whac-a-mole-that-web-giants-just-wont-win-154427">been</a> <a href="http://theconversation.com/does-deplatforming-work-to-curb-hate-speech-and-calls-for-violence-3-experts-in-online-communications-weigh-in-153177">widely</a> <a href="http://theconversation.com/deplatforming-online-extremists-reduces-their-followers-but-theres-a-price-188674">debated</a>, as the intervention treads a thin line between preventing harm and censoring speech. How to weigh the benefits and harms of these practices? How to assess whether they are effective? Enter research on content curation.</p><p>Despite this promise, however, research on content curation practices has arguably failed to drive their development and adjustment. Compared to other (polarizing) topics like healthcare or labor, content curation practices in social media are disproportionally driven by anecdotal or observational evidence that describes problems, but not solutions. This is (at least partly) because researching content curation practices is damn hard:</p><ul><li><p>Content curation practices are opaque and carried by private companies at their discretion. </p></li><li><p>Researchers often lack access to data or the necessary experimentation infrastructure; </p></li><li><p>Online platforms are highly dynamic, raising concerns about the <a href="https://kevinmunger.substack.com/p/temporal-validity-is-distinct-from">temporal validity of findings</a>; </p></li><li><p>And, in some cases, disentangling the effect of content curation practices is methodologically challenging, e.g., <a href="https://dl.acm.org/doi/abs/10.1145/3351095.3372879">see</a> <a href="https://www.pnas.org/doi/abs/10.1073/pnas.2101967118">the</a> <a href="https://www.pnas.org/doi/10.1073/pnas.2101967118">literature</a> on the effects of the YouTube algorithm.</p></li></ul><p>Yet, we should not give up! Causal evidence on content curation practices is challenging but can yield significant payoffs: given how widely used online platforms are, marginal improvements to online environments are meaningful; and given that there's a wide appetite for regulating online platforms, research can guide policy away from guesswork.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://doomscrollingbabel.manoel.xyz/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Doomscrolling Babel! Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>