article - ai and the problem of knowledge collapse
article - ai and the problem of knowledge collapse
Andrew J. Peterson
April 2024
arXiv:2404.03502v2 [cs.AI] 22 Apr 2024
While artificial intelligence has the potential to process vast amounts of data, generate new insights,
and unlock greater productivity, its widespread adoption may entail unforeseen consequences. We
identify conditions under which AI, by reducing the cost of access to certain modes of knowledge,
can paradoxically harm public understanding. While large language models are trained on vast
amounts of diverse data, they naturally generate output towards the ‘center’ of the distribution.
This is generally useful, but widespread reliance on recursive AI systems could lead to a process we
define as “knowledge collapse”, and argue this could harm innovation and the richness of human
understanding and culture. However, unlike AI models that cannot choose what data they are
trained on, humans may strategically seek out diverse forms of knowledge if they perceive them to
be worthwhile. To investigate this, we provide a simple model in which a community of learners or
innovators choose to use traditional methods or to rely on a discounted AI-assisted process and
identify conditions under which knowledge collapse occurs. In our default model, a 20% discount
on AI-generated content generates public beliefs 2.3 times further from the truth than when there
is no discount. An empirical approach to measuring the distribution of LLM outputs is provided in
theoretical terms and illustrated through a specific example comparing the diversity of outputs
across different models and prompting styles. Finally, we consider further research directions to
counteract harmful outcomes.
Before the advent of generative AI, all text and artwork was produced by humans, in some cases
aided by tools or computer systems. The capability of large language models (LLMs) to generate
text with near-zero human effort, however, along with models to generate images, audio, and video,
suggest that the data to which humans are exposed may come to be dominated by AI-generated or
AI-aided processes.
Researchers have noted that the recursive training of AI models on synthetic text may lead to
degeneration, known as “model collapse” (Shumailov et al. 2023). Our interest is in the inverse
of this concern, focusing instead on the equilibrium effects on the distribution of knowledge
within human society. We ask under what conditions the rise of AI-generated content and AI-
mediated access to information might harm the future of human thought, information-seeking,
and knowledge.
The initial effect of AI-generated information is presumably limited, and existing work on the
harms of AI rightly focuses on the immediate effects of false information spread by “deepfakes”
(Heidari et al. 2023), bias in AI algorithms (Nazer et al. 2023), and political misinformation (Chen
and Shu 2023). Our focus has a somewhat longer time horizon, and probes the impact of widespread,
rather than marginal adoption.
Researchers and engineers are currently building a variety of systems whereby AI would mediate
our experience with other humans and with information sources. These range from learning from
LLMs (Chen, Chen, and Lin 2020), ranking or summarizing search results with LLMs (Sharma, Liao,
and Xiao 2024), suggesting search terms or words to write as with traditional autocomplete (Graham
2023; Chonka, Diepeveen, and Haile 2023), designing systems to pair collaborators (Ball and Lewis
2018), LLM-based completion of knowledge bases sourced from Wikipedia (Chen, Razniewski, and
Weikum 2023), interpreting government data (Fisher 2024) and aiding journalists (Opdahl et al.
2023), to cite only a few from an ever-growing list.
Over time, dependence on these systems, and the existence of multifaceted interactions among
them, may create a “curse of recursion” (Shumailov et al. 2023), in which our access to the original
diversity of human knowledge is increasingly mediated by a partial and increasingly narrow subset
of views. With increasing integration of LLM-based systems, certain popular sources or beliefs
which were common in the training data may come to be reinforced in the public mindset (and
within the training data), while other “long-tail” ideas are neglected and eventually forgotten.
Such a process might be reinforced by an ‘echo chamber’ or information cascade effect, in
which repeated exposure to this restricted set of information leads individuals to believe that
the neglected, unobserved tails of knowledge are of little value. To the extent AI can radically
discount the cost of access to certain kinds of information, it may further generate harm through
the “streetlight effect”, in which a disproportionate amount of search is done under the lighted area
not because it is more likely to contain one’s keys but because it’s easier to look there. We argue
that the resulting curtailment of the tails of human knowledge would have significant effects on
a range of concerns, including fairness, inclusion of diversity, lost-gains in innovation, and the
preservation of the heritage of human culture.
In our simulation model, however, we also consider the possibility that humans are strategic
in actively curating their information sources. If, as we argue, there is significant value in the tai’
areas of knowledge that come to be neglected by AI-generated content, some individuals may put
in additional effort to realize the gains, assuming they are sufficiently informed about the potential
value.
We identify a dynamic whereby AI, despite only reducing the cost of access to certain kinds
of information, may lead to “knowledge collapse,” neglecting the long-tails of knowledge and
creating an degenerately narrow perspective over generations. We provide a positive knowledge
spillovers model with in which individuals decide whether to rely on cheaper AI technology or
invest in samples from the full distribution of true knowledge. We examine through simulations
the conditions under which individuals are sufficiently informed to prevent knowledge collapse
within society. To evaluate this empirically, we outline an approach to defining and measuring
output diversity, and provide an illustrative example. Finally, we conclude with an overview of
possible solutions to prevent knowledge collapse in the AI-era.
2. Previous Work
Technology has long affected how we access knowledge, raising concerns about its impact on
the transmission and creation of knowledge. Yeh Meng-te, for example, argued in the twelfth
century that the rise of books led to a decline in the practice of memorizing and collating texts that
contributed to a decline of scholarship and the repetition of errors (Cherniack 1994). Even earlier, a
discussion in Plato’s Phaedrus considers whether the transition from oral tradition to reading texts
was harmful to memory, reflection and wisdom (Hackforth 1972).
We focus on recent work on the role of digital platforms and social interactions, and mention
only in passing the literature on historical innovations and knowledge (Ong 2013; Mokyr 2011;
Havelock 2019), and the vast literature on the printing press (Dittmar 2011; Eisenstein 1980). Like
other media transitions before it (Wu 2011), the rise of internet search algorithms and of social
media raised concerns about the nature and distribution of information people are exposed to, and
the downstream effects on attitudes and political polarization (Cinelli et al. 2021; Barberá 2020).
The following section considers research on the impact of recommendation algorithms and
self-selection on social media, and how this might generate distorted and polarizing opinions, as an
analogy for understanding the transformation brought about by reliance on AI. We consider game
theoretic models of information cascades as an alternative model for failure in social learning,
in which the public to fails to update rationally on individuals’ private signals. Next, we review
the main findings of network analysis on the flow of information in social media, which also
identify mechanisms which distort knowledge formation. We then examine the specific nature
of generative AI algorithms, focusing on the problem of model collapse and known biases in AI
outputs.
A common critique of social media is that they allow users to select in to “echo chambers” (specific
communities or communication practices) in which they are exposed to only a narrow range of
topics or perspectives. For example, instead of consulting the “mainstream” news where a centrist
and relatively balanced perspective is provided, users are exposed to selective content that echoes
pre-existing beliefs. In the ideological version of the echo-chamber hypothesis, individuals within
a latent ideological space (for example a one-dimensional left-right spectrum), are exposed to peers
and content with ideologically-similar views. If so, their beliefs are reinforced socially and by a
generalization from their bounded observations, leading to political polarization (Cinus et al. 2022;
Jamieson and Cappella 2008; Pariser 2011).
A simple model for this assumes homophily within in a network growth model, in which similar
individuals chose to interact. Implicitly the approach presumes that this is common on social
media but not common within traditional media, which for technological reasons were constrained
to provide the same content across a broad population with possibly heterogeneous preferences.1
This general dynamic may hold even if traditional media and newspapers were themselves dynamic
systems interacting with their consumers, markets and advertisers, and themselves adapting their
message to specific communities and preferences (Angelucci, Cagé, and Sinkinson forthcoming;
Cagé 2020; Boone, Carroll, and van Witteloostuijn 2002) .
The second main line of analysis focuses on “filter bubbles,” whereby the content to which users
are exposed is selected based on a recommendation system. Jiang, et al., model this as a dynamic
process between a user’s evolving interests and behavior (such as clicking a link, video, or text)
1
The reality is as usual more complex. For example, in the post-war era, the concern was almost the inverse- the fear
that the few channels that were possible with television led to ‘homogenization.’ There are also other dynamics at play
than technological constraints. For example, in contrast to TV, the 1950s and 1960s saw a proliferation of more diverse
and local radio stations, some catering to ethnic minorities and musical tastes outside the mainstream. The ‘payola’
scandals, however, led to regulations that shifted content decisions from diverse DJs to centralized music directors
(Douglas 2002).
and a recommender system which aims to maximize expected utility for the user (Jiang et al. 2019).
In their reinforcement learning-inspired framework, the aim is for the user to explore the space of
items or topics without the algorithm assigning degenerate (extremely high or zero) probabilities
to these items. As above, a key concern is the political or ideological content of recommendations
their relation to polarization (Keijzer and Mäs 2022). In a more recent twist, (Sharma, Liao, and Xiao
2024) find that LLM-powered search may generate more selective exposure bias and polarization
by reinforcing pre-existing opinions based on finer-grained clues in the user’s queries.
Particularly relevant for our context is the issue of “popularity bias” in recommender systems, in
which a small subset of content receives wide exposure while users (distributed based on some long-
tailed distribution, like the topics) from smaller groups or with rare preferences are marginalized.
On the one hand, users may desire to be exposed to popular content, for example to understand
trending ideas or fashions. But overly favoring popular items can lead to user disengagement
because it neglects their unique interests, lacks variety, etc. (Klug et al. 2021) . Recommendation
systems are often biased in the sense that even when a subset of users wants to get access to
non-popular items, they receive few or no such recommendations. A number of approaches have
been suggested to counteract this tendency (Lin et al. 2022; Gao et al. 2023).
The problem of popularity bias is ironic given that one of the unique contributions of the
internet was its ability to provide access to long-tailed products and services that were previously
ignored or inaccessible (Brynjolfsson, Hu, and Smith 2006, 2003). By extension, we would expect
social media and the internet to make possible a more diverse and rich informational environment.
The role of self-selection into communities and recommendation algorithms provides a explanation
for why this might not be the case. In the next section we consider a more general set of models
that examine information flow within networks and the idea of information cascades.
Information cascade models provide one approach to explaining a kind of herd behavior (where
diverse and free individuals nonetheless make similar decisions). They explore the conditions
under which private information is not efficiently aggregated by the public. This can occur where
individuals sequentially make decisions from a discrete set after observing the behaviors but not
the private signals of others. This can generate a “herd externality” (Banerjee 1992) in which an
individual ignores her private signal in deciding, and as a result the public is in turn unable to
update on her private information. In the extreme, this can mean that all private information, aside
from that of the first few individuals, is completely ignored (Bikhchandani, Hirshleifer, and Welch
1998; Smith and Sørensen 2000). In some variants of the model, individuals must pay to receive a
signal, which encourages the tendency to want to free-ride on the information received by others,
and thus the greater the cost, the more likely it is that a cascade develops.
A related literature on the spread of information on social networks analyzes information
cascades in terms of network structure, as a kind of contagion. Here, the focus is not on private
information but how information flows within the network. For example, independent cascade
models consider how an individual may change their beliefs based on some diffusion probability
as a result of contact with a neighbor with that belief (Goldenberg, Libai, and Muller 2001; Gruhl et
al. 2004).
More generally, such models determine the probability of diffusion within a network as some
function of the connected nodes, and may also incorporate additional characteristics such as
each nodes’ social influence, ideological or other preferences, or topics (Barbieri, Bonchi, and
Manco 2013). Alternatively, epidemic models allow that individuals may be in one of three states -
susceptible, infected (capable of transmitting the information), and recovered (in which case they
have the information but do not consider it worth sharing with others) (Kermack and McKendrick
1927; Barrat, Barthélemy, and Vespignani 2008) .
Social (and even physical) proximity can lead individuals to share similar attitudes, such as when
individuals randomly assigned housing together come to have attitudes similar to their apartment
block and differing from nearby blocks (Festinger, Schachter, and Back 1950; Nowak, Szamrej,
and Latané 1990). Empirically, some claim that weak-ties may be more important for information
diffusion that strong-ties (Bakshy et al. 2012), while another approach focuses on clustering within
the network as a means for spreading information (Centola 2010). More sophisticated models
allow for the evolution not only of opinion process but the edges between nodes of the network
(Castellano, Fortunato, and Loreto 2009).
These models suggest specific opinion-formation dynamics based on what other humans, texts,
images, etc. an individual interacts with. By extension, we could consider the generalization of these
networks to the case where LLMs play a key role as (possibly influential) nodes, or as determining
how an individual navigates a knowledge graph. One of the key ideas of Web 2.0 was that users, not
just authors or programmers, structure the knowledge (O’Reilly 2005). By extension, in the AI era,
LLMs interact with users, authors, programmers and technology to structure that knowledge, and
understanding the flow of information requires understanding the emergent behavior of these
elements.
The idea of model collapse is rooted in the earlier phenomenon of “mode collapse” in generative
adversarial networks (GANs). GANs are based on a generator neural network that proposes, e.g. an
image, and a discriminator attempts to predict whether a given image is created by the generator
or is a real image from the dataset. While ideally the generator attempts to produce images across
the full range of input data, in practice they may settle into producing a narrow range of images
for which it is good at fooling the discriminator, known as mode collapse (Goodfellow 2016; Arora
et al. 2017). The case of “posterior collapse” was also identified in modeling language data with
variational autoencoders (Melis, György, and Blunsom 2022).
Shumailov, et al. (2023) introduced the term “model collapse” to describe a related process when
models such as variational autoencoders, Gaussian mixture models, and LLMs are trained on data
produced by an earlier version of the model. Incorporating AI-generated content in the training
data causes loss of information which they categorize into two types. First, in “early model collapse,”
the tails of the distribution are lost due to statistical error (finite sampling bias) or functional
approximation error, which leads to reversion to the mean. Second, “late model collapse” may
occur when a model converges with narrow variance on a distribution unlike the original data.
They provide evidence of such model collapse in LLMs and other models, see for example Figure 1.
Input: some started before 1360 — was typically accomplished by a master mason and a small team of itinerant
masons, supplemented by local parish labourers, according to Poyntz Wright. But other authors reject
this model, suggesting instead that leading architects designed the parish church towers based on early
examples of Perpendicular
Outputs:
Gen 0: Revival architecture such as St. John’s Cathedral in London. The earliest surviving example of Perpen-
dicular Revival architecture is found in the 18th @-@ century Church of Our Lady of Guernsey, which
dates from the late 19th century. There are two types of perpendicular churches : those
Gen 1: architecture such as St. Peter’s Basilica in Rome or St. Peter’s Basilica in Buenos Aires. There is no
evidence that any of these buildings were built during the reign of Pope Innocent III, but it is possible
that they may have been built during the reign of his successor, Pope Innocent
Gen 7: architecture in England. In an interview with The New York Times, Wright said : " I don ’t think there is
anything wrong with me being able to do what I want to do. It just doesn ’t work for me. " He added : " I
don ’t know if you can call it funny,
Gen 9: architecture. In addition to being home to some of the world’s largest populations of black @-@ tailed
jackrabbits, white @-@ tailed jackrabbits, blue @-@ tailed jackrabbits, red @-@ tailed jackrabbits, yellow
@-
On the one hand, under some conditions the injection of true (non AI-generated) data can
preserve representation of the true distribution (Dohmatob et al. 2024), though in other models,
small amounts of synthetic data can poison an image model, and once distorted, it is difficult for
such models to recover even after being trained on true data (Bohacek and Farid 2023). As one
particular example, training LLMs on synthetic data can lead to diminishing lexical, semantic and
syntactic diversity (Guo et al. 2023).
2.4. Known biases in LLMs
Newer AI models such as LLMs are not immune to the problems of bias identified and measured in
machine learning algorithms (Nazer et al. 2023) and which have plagued predictive algorithms in
real-world uses cases going back to at least the 1930s (Christian 2021). Unsurprisingly, LLMs are
better at recalling facts that occur frequently within the training data and struggle with long-tail
knowledge (Kandpal et al. 2023). (Das et al. 2024) identify a range of shortcomings of LLMs in
attempting to generate human-like texts, such as under-representing minority viewpoints and
reducing the broad concept of “positive” text to that simply of expressing “joy”.
Recent work attempts to address these issues through a variety of methods, for example by
upsampling under-represented features on which prediction is otherwise sub-optimal (Gesi et
al. 2023), or evaluating the importance of input data using shapely values (Karlaš et al. 2022).
However, the mechanistic interpretability work on LLMs to date suggest that our understanding,
while improving, is still very limited (Kramár et al. 2024; Wu et al. 2023). As such, direct methods
for overcoming such biases are, at a minimum, not close at hand. Finally, while much of the focus is
naturally on overt racial and gender biases, there may also be pervasive but less observable biases
in the content and form of the output. For example, current LLMs trained on large amounts of
English text may ‘rely on’ English in their latent representations, as if a kind of reference language
(Wendler et al. 2024).
One particular area in which the diversity of LLM outputs has been analyzed is on a token-by-
token level in the context of decoding strategies. In some situations, using beam search to choose
the most likely next token can create degenerate repetitive phrases (Su et al. 2022). Furthermore, a
bit like Thelonious Monk’s melodic lines, humans do not string together sequences of the most
likely words but occasionally try to surprise the listener by sampling from low-probability words,
defying conventions, etc. (Holtzman et al. 2020).
A commonly held, optimistic view is that knowledge has improved monotonically over time, and
will continue to do so. This indeed appears to be the case for certain scientific fields like physics,
chemistry, or molecular biology, where we can measure the quality of predictions made over time.
For example, accuracy in the computation of digits of π has increased from 1 digit in 200 BCE to 16
in 1424 (Jamashid al-Kashi) to 1014 digits recently.
In other domains, however, it is less clear, especially within regions. Historically, knowledge
has not progressed monotonically, as evidenced by the fall of the Western Roman empire, the
destruction of the House of Wisdom in Baghdad and subsequent decline of the Abbasid Empire after
1258, or the collapse of the Mayan civilization in the 8th or 9th century. Or, to cite specific examples,
the ancient Romans had a recipe for concrete that was subsequently lost, and despite progress
we have not yet re-discovered the secrets of its durability (Seymour et al. 2023), and similarly for
Damascus steel (Kürnsteiner et al. 2020). Culturally, there are many languages, cultural and artistic
practices, and religious beliefs that were once held by communities of humans which are now lost
in that they do not exist among any known sources (Nettle and Romaine 2000).
The distribution of knowledge across individuals also varies over time. For example, traditional
hunter-gatherers could identify thousands of different plants and knew their medicinal usages,
whereas most humans today only know a few dozen plants and whether they can be purchased
in a grocery store. This could be seen as a more efficient form of specialization of information
across individuals, but it might also impact our beliefs about the value of those species or of a walk
through a forest, or influence scientific or policy-relevant judgements.
Informally,2 we define knowledge collapse as the progressive narrowing over time (or over
technological representations) of the set of information available to humans, along with a con-
comitant narrowing in the perceived availability and utility of different sets of information. The
latter is important because for many purposes it is not sufficient for their to exist a capability to, for
example, go to an archive to look up some information. If all members deem it too costly or not
worthwhile to seek out some information, that theoretically available information is neglected and
useless.
The main focus of the model is whether individuals decide to invest in innovation or learning
(we treat these as interchangeable) in the ‘traditional’ way, through a possibly cheaper AI-enabled
process, or not at all. The idea is to capture, for example, the difference between someone who
does extensive research in an archive rather than just relying on readily-available materials, or
someone who takes the time to read a full book rather than reading a two-paragraph LLM-generated
summary.
Humans, unlike LLMs trained by researchers, have agency in deciding among possible inputs.
Thus, a key dynamic of the model is to allow for the possibility that rational agents may be able to
prevent or to correct for distortion from over-dependence on ‘centrist’ information. If past samples
neglect the ‘tail’ regions, the returns from such knowledge should be relatively higher. To the extent
that they observe this, individuals would be willing to pay more (put in more labor) to profit from
these additional gains. We thus investigate under what conditions such updating among individuals
is sufficient to preserve an accurate vision of the truth for the community as a whole.
2
For further discussion and a more precise definition using the notation from the model, see the Appendix.
The cost-benefit decision to invest in new information depends on the expected value of that
information. Anyone who experiments with AI for, e.g. text summarization, develops an intuitive
sense of when the AI provides the main idea sufficiently well for a given purpose and when it is
worth going straight to the source. We assume that individuals cannot foresee the future, but they
do observe in common the realized rewards from previous rounds. The decision also depends on
each individual’s type. Specifically, n individuals have types Θn drawn from a lognormal distribution
with µ = 1, σ = 0.5. Depending on how their utility is calculated (not a substantive focus here), these
could be interpreted as different expected returns from innovation (e.g.techno-optimists versus
pessimists), or their relative ability or desire to engage in innovation.
Notation Description
n number of individuals (= 25)
Θn the type of individual n, multiplying their expected return from innovation
d. f . degrees of freedom for t-distribution, determines width of the tails
ptrue (x) the ‘true’ probability distribution function, t-distribution with e.g. 10 d.f.
ppublic (x) the public approximation to the true pdf based on the 100 most
recent samples using kernel density estimation
δ AI-discount rate, where the cost of an AI-sample is δ times
the cost of a sample from the full distribution
σtr truncation limits for the AI-generated samples, in standard deviations
v̂t The estimated value of a sample at time t
η learning rate, i.e. how quickly individuals update
their beliefs on the value of full and truncated
samples based on samples observed in the last round
r How many rounds between generations
(if greater than 100, no generational effects)
I Innovation from an individual’s sample
(i.e. how far they move the public pdf towards the true pdf)
determines the n’s payout when multiplied by Θn
In Figure 2, we illustrate the innovation calculation for a hypothetical example where the
distance between the existing public pdf and the true pdf is 0.5, while the n + 1th sample reduces
the distance to 0.4, thereby generating an innovation of 0.1.
This can be thought of as akin to a patent process, in which an individual receives rents for her
patent (to the extent that it is truly innovative) in exchange for contributing to public knowledge
that benefits others.
As noted above, individuals cannot foresee the true future value of their innovation options
(they do not know what sample they will receive or how much value it will add. Instead, they can
only estimate the relative values of innovation based on the previous rounds. Specifically, they
update their belief about the options based on the previous full and truncated (AI) samples from
the previous round (and a minimum of three), according to a learning rate (η) as follows. For the
previous estimate v̂t−1 , the new estimate v̂t for each of the full- and truncated-samples is calculated
4
Varying this has trivial effect on the model, though higher values can distort public knowledge.
5
We use the Hellinger distance because it is a true distance metric that is symmetric and satisfies the triangle
inequality, which is important for the innovation calculation. The Hellinger distance is bounded by 0 and 1 (if the two
pdfs have no √ common support) and given by:
√ √
H( p, q) = √1
2
∫ ( p(x) − q(x)) dx
2
pnpublic
∆ new sample
0.5
0.4 ptrue
pn+1
public
I = 0.5 − 0.4 = 0.1
FIGURE 2. A hypothetical innovation calculation where the new (n+1)th sample moves the public
pdf 0.1 towards the true distribution.
By varying the learning rate, we can evaluate the impact of having more or less up-to-date in-
formation on the value of different information sources, where we expect that if individuals are
sufficiently informed, they will avoid knowledge collapse by seeing and acting on the potential to
exploit knowledge from the tail regions, even if relatively more expensive.
While the individual payoff is based on the true movement of the public pdf towards the true
pdf, the public pdf is updated based on all samples. This reflects that public consciousness is
overwhelmed with knowledge claims and cannot evaluate each, so that a consensus is formed
around the sum of all voices. Unlike the individual innovator who has a narrow focus and observes
whether her patent ultimately generates value, the public sphere has limited attention and is forced
to accept the aggregate contributions of the marketplace of ideas.
As a result, individuals’ investments in innovation have positive spillovers to the extent they
can move public knowledge towards the truth. However, if too many people invest in ‘popular’ or
‘central’ knowledge by sampling from the truncated distribution, this can have a negative externality,
by distorting public knowledge towards the center and thinning the tails.6
We also introduce the possibility of generational turnover in some models to explore the
impact on knowledge collapse. This could either be taken to be literal generations of humans, as in
6
If individuals knew they were sampling from a truncated distribution, they could use the Expectation-Maximization
algorithm to recover the full distribution, but again this process is meant to be metaphorical, and there is no known
real-life method for recovering the source knowledge from AI-generated content.
economic ‘overlapping generation’ models (Weil 2008), or alternatively as reflecting the recursive
nature of reliance on interleaved AI-systems, which could generate the same result within a rapid
timeframe.
In the version of the model with generational change, the new generation takes the existing
public pdf to be representative and thus begins sampling from a distribution with the same (possibly
smaller) variance (and correspondingly the truncation limits are updated). Interpreted in terms of
human generations, this could be understood as the new generation fixing its ‘epistemic horizon’
based on the previous generation. That is, the new generation may underestimate the breadth of
possible knowledge and then rely on these perceived limits to restrict their search.7 An information
cascade model could justify such a situation if individuals assume that previous actors would
have invested in tail knowledge if was valuable, and thus take the absence of such information as
implying that it must be of little value.8
A second interpretation views these ‘generations’ not in terms of human populations but as
a result of recursive dynamics among AI systems, such as when a user reads an AI-generated
summary of an AI-written research article which was itself constructed from Wikipedia articles
edited with AI, etc., a fancy version of the telephone game.
4. Results
Our main concern is with the view that AI, by reducing the costs of access to certain kinds of
information, could only make us better off. In contrast to the literature on model collapse, we
consider the conditions under which strategic humans may seek out the input data that will
maintain the full distribution of knowledge. Thus, we begin with a consideration of different
discount rates. First, we present the a kernel density estimate of public knowledge at the end
of 100 rounds (Figure 3). As a baseline, when there is no discount from using AI (discount rate
is 1), then as expected public knowledge converges to the true distribution,9 As AI reduces the
cost of truncated knowledge, however, the distribution of public knowledge collapses towards the
center, with tail knowledge being under-represented. Under these conditions, excessive reliance
7
Zamora (2010) suggests a scientific process of ‘verisimiltude’, where we judge evidence not with reference to objective
truth by by “perceived closeness to what we empirically know about the truth, weighted by the perceived amount of
information this empirical knowledge contains” (Zamora-Bonilla 2010) . Work on human cultural transmission attempts
to explain, for example, how the Tasmanians lost a number of useful technologies over time. (Mesoudi and Whiten 2008;
Henrich 2004)
8
For example, Christian communities at times actively promoted and preserved ‘canonical’ texts while neglecting
or banning others, with the result that those excluded from reproduction by scribes were taken to have little value.
Perhaps the heliocentric view espoused by Aristarchus of Samos in the 3rd century BCE would have been more readily
(re)considered if his works had not been neglected (Russo and others 2003). A number of authors, such as Basilides, are
known to us today only through texts denouncing (and sometimes misrepresenting) their views (Layton 1989).
9
Even with no discount, there are occasional samples from the truncated distribution, but only enough to realize that
they are of relatively less worth than full-distribution samples
on AI-generated content over time leads to a curtailing of the eccentric and rare viewpoints that
maintain a comprehensive vision of the world.
0.6
0.4
0.2
0.0
1.5 1.0 0.5 0.0 0.5 1.0 1.5
FIGURE 3. Knowledge collapse: The cheaper it is to rely on AI-generated content, the more extreme
the degeneration of public knowledge towards the center.
Fixing specific parameters, we can get a sense of the size of the the impact of relying on AI. For
example, for our default model,10 after nine generations, when there is no AI discount the public
distribution has a Hellinger distance of just 0.09 from the true distribution11 . When AI-generated
content is 20% cheaper (discount rate is 0.8), the distance increases to 0.22, while a 50% discount
increases the distance to 0.40. Thus, while the availability of cheap AI-approximations might be
thought to only increase public knowledge, under these conditions public knowledge is 2.3 or 3.2
times further away from the truth due to reliance on AI.
For subsequent results illustrating the tradeoff of different parameters, we plot the Hellinger
distance between public knowledge at the end of the 100 rounds and the true distribution. First, we
examine the importance of updating on the value of relative samples and the relationship to the
discount factor in Figure 4. That is, we compare the situation in which individuals do not update on
the value of innovation in previous rounds (learning rate near zero, e.g. η = 0.001) to the case where
they update rapidly (here η = 0.1). As above, the more AI-generated content is cheaper (discount
rate indicated by colors), the more public knowledge collapses towards the center. At the same
time, when individuals update more slowly on the relative value of learning from AI (the further
to the left in the figure), the more public knowledge collapses. We also observe a tradeoff, that is,
faster updating on the relative value of AI-generated content can compensate for more extreme
10
Truncation at σt r = 0.75 standard deviations from the mean, generations every 10 rounds, learning rate of 0.05.
11
Even here there are occasional samples from the truncated distribution – just enough to realize that they have less
value than the full-distribution samples.
0.55 Discount Factor
0.1
0.2
0.50 0.3
0.4
0.5
0.45 0.6
0.7
0.8
0.9
0.40 1.0
Hellinger Distance
0.35
0.30
0.25
0.20
0.15
0.00 0.02 0.04 0.06 0.08 0.10
Learning Rate (lr)
price disparities. And conversely, if the discount rate is not too extreme, even slower updating on
the relative values is not too harmful.
In Figure 5, we consider the impact of variations in how extreme the truncation of AI-generated
content is on the collapse of knowledge. Intuitively, extreme truncation (small values of σtr ) corre-
spond to a situation in which AI, for example, summaries an idea with only the most obvious or
common perspective. Less extreme truncation corresponds to the idea that AI manages to represent
a variety of perspectives, and excludes only extremely rare or arcane perspectives. Naturally, in
the latter case, (e.g. if AI truncates the distribution two standard deviations from the mean), the
effect is minimal. If AI truncates knowledge outside of 0.25 standard deviations from the mean, the
impact is large, though once again this is at least someone moderated when the discount is smaller
(especially if there is no generational effect).
We compare the effect of the generational compounding of errors in Figure 6. If there is no
generational change, there is at worst only a reduction in the tails of public knowledge outside the
truncation limits. In this case the distribution is stable and does not “collapse”, that is, over time
the problem is not progressively worse. We see a jump from this baseline to the case where there is
generational change, though the effect of how often generational change occurs (every 3, 5, 10, or
20 rounds) does not have a significant impact.
Discount Factor
0.1
0.2
0.6 0.3
0.4
0.5
0.6
0.7
0.8
0.5 0.9
1.0
Hellinger Distance
0.4
0.3
0.2
0.4
0.3
0.2
While the above discussion focused on the effects of medium- to long-term adoption, we provide
a brief theoretical overview of approaches to measuring knowledge collapse. We also illustrate
measuring the distribution of LLM output in line with a specific empirical example, and provide
a brief analysis of the extent to which direct prompting can help generate more diverse outputs.
Unlike typical evaluations of machine learning algorithms, our focus is not on an evaluation metric
relative to a set of gold-standard input- output pairs x and y, but the diversity of outputs across a
distribution of possible responses for a given question or prompt. For our illustration, we consider
the question of “What does the well-being of a human depend on?”, as a well-known question
on which there no general consensus on a particular answer but instead every culture generally
provides an answer in more or less explicit terms.
We consider situations in which an LLM generates as outputs a set of items out of some possibly-
larger set of possibilities. For example, we might ask the LLM to (repeatedly) produce a list of 100
human languages, poets, animal species, or chemical compounds.
The LLM generates this set based on a transformer, state-space model, etc. M, based on a
set of parameters θ by generating, e.g. an auto-regressive set of tokens. A strong definition of
representativeness would say that the probability of inclusion of any member xi in the output is
equal to the probability of inclusion within a random sample from the full set X.
1
(1) ∀xi , x j ∈ X. PM(θ) (xi ) = PM(θ) (x j ) =
∣X∣
where PM(θ) (xi ) denotes the probability of item xi being included in the output distribution generated by
M with parameters θ, and ∣X∣ is the cardinality of set X.
While useful in some particular cases, or in the absence of alternatives, this is unreasonable for
many real-world use cases. The other extreme is to introduce a simple non-degeneracy condition,
saying that any item should have at least some chance of appearing in the LLM output:
DEFINITION 2 (MINIMAL REPRESENTATIVENESS). Minimal Representativeness requires that for any
item xi ∈ X, the probability of xi being included in the output distribution generated by M with any given
parameters θ, is non-zero, such that
where PM(θ) (xi ) denotes the probability of item xi being included in the output distribution generated by
M.
Intuitively, if asking an LLM for a diverse list of items, any possible item should have some non-
zero probability of inclusion, analogous to the definition of non-degeneracy in a recommendation
algorithm (Jiang et al. 2019).
In some cases another possible practical approach is to define representativeness relative to a
specific task. This implicitly defines a set of values across the different outcomes. For example, we
might want the likelihood of a language of appearing in the output to be proportional to the number
of currently-living speakers of that language. For some purposes, this would be the most helpful
output, although in other cases, such as for a historian, it might be problematic because it fails to
consider languages with no living speakers. Other users, for example, might want to consider the
possibility of up-weighting at-risk languages in hopes of preserving them, etc.
w(xi )
(3) PM(θ) (xi ) =
∑x j ∈X w(x j )
where PM(θ) (xi ) denotes the probability of item xi being included in the output distribution generated by
M.
For example, a doctor asking for a list of possible infectious diseases to diagnose a patient might
expect that it would begin with the most commonly occurring diseases first, before mentioning
more obscure ones. Even this list depends on context, as there are certain infectious diseases are
common worldwide (e.g. tuberculosis) but are rare in certain countries.
Finally, a similar measure considers the extent to which different (e.g. protected) groups are
represented in the output. As an example, in a small corpus created by asking Phi-2 2.7B and
Llama2-13b to list twenty languages 100 times each, there is only one single mention of a non-verbal
(Sign) language (“German Sign Language”), despite there being around 72 million people worldwide
who speak sign languages. By contrast, Italian is spoken by around 68 million people, which is
slightly less, but Italian was mentioned 508 times. There are obvious reasons for this, in terms
of the frequency of appearance of references to Italian in the text, but it is worth considering
the possibility that this might reinforce biases and narrowness in thinking about languages. An
objection might be that Italian is one language while the speakers of Sign languages use more than
300 different languages. Still, it’s worth noting that Italian is 29th in the world in terms of number of
speakers, so the list of twenty was not representative of the top twenty languages by speakers either,
neglecting Tagalog 64 mentions; 83 million speakers, Tamil (58 mentions; 87 million speakers), etc.
This suggests the value of a general measure of this:
w(Gi )
(4) PM(θ) (x ∈ Gi ) =
k
∑ j=1 w(G j )
for all x ∈ Gi and for each group Gi , where PM(θ) (x ∈ Gi ) denotes the probability of an individual from
group Gi being included in the output distribution generated by M.
This definition does not take into account the problems associated with categorizing people into
such groups or the effect of intersectionality, but these categories function as reference points as
they commonly serve as a basis for research, public discourse, and policy decisions.
Version Prompt
The entities from each response for prompts 1-3 were extracted by asking GPT-3.5-turbo to
extract a list of entities from the text, while for prompts 4 and 5 these were parsed directly from
the results. One specific challenge is to map all outputs to a unique but comprehensive list of
entities. Our approach consisted of first generating a large set of candidate entities, then doing
entity resolution using embeddings and DBSCAN algorithm (Schubert et al. 2017), as follows. First,
the corpus of possible entities came from (1) the list of entities generated by the LLMs, and (2)
extracting entities from the Internet Encyclopedia of Philosophy (Fieser and Dowden 2024) and
then identifying those which are relevant to the question based on a query to GPT-3.5-turbo. An
additional pass through the data asking GPT-3.5-turbo for the associated regions was the basis
for the region entities mentioned above. The embedding for each entity was generated using the
HuggingFace ‘all-MiniLM-L6-v2’ model, and similar identities identified using DBSCAN with ϵ = 3.0.
Identities which were not matched to a cluster based on this cutoff were discarded, meaning that
there is a possibly still-longer-tail of answers which is not measured by this method. This generated
the final ‘reference’ list of 2,693 entities, to which any entity identified in an output was mapped,
again using the same embeddings model.
Each step of this process introduces possible errors, in that (a) the the reference list does
not include all possible entities, and is likely biased against less mainstream philosophers,14 (b)
the reference list may contain some entities that are not clearly good answers to the question
(for example, being philosophers who did not make significant contributions to ethics/ morality).
Finally, (c) biased could be introduced by the mapping of entities to the reference list, for example
because of multiple names used for the same entity15 or names that are close in the embedding
14
Even the Internet Encyclopedia of Philosophy, while including diverse philosophy from all over the world, includes
a clear focus on Anglo-American and European philosophy, reflecting the interest of its writers and editors.
15
For example, the full name for “Avicenna” is actually “Abū Álı̄ al-H
. usayn bin Ábdullāh ibn al-H
. asan bin Álı̄ bin Sı̄nā
space but nonetheless distinct entities. For the comparative purposes of the task, however, it is no
clear reason to expect that these biases systematically distort the outcomes.
We analyze the results through three approaches. First, we calculate Shannon’s Diversity Index (H ′ )
a version of Shannon entropy based on the natural logarithm that is commonly used in ecological
studies to measure diversity within a community:
R
H ′ = − ∑ pi ln(pi )
i=1
where R represents the total number of unique entities in the dataset (defined by the 2,693 reference
entities), and pi is the proportion of the total population for entity i. We also calculate Pielou’s
Evenness Index (J ′ ), which normalizes Shannon’s diversity index relative to a maximum possible
H ′ that would occur if all entities were equally-well represented (Pielou 1966), corresponding to
Definition 1 above.
The results are reported in Table 3. There does not appear to be significant variation in the
different prompts, with the exception of prompt 5, which does generate significantly more diverse
responses (mean J ′ of 0.86 compared to 0.55 for prompt 1). This is represented visually in Figure 7),
in which prompt version 5, which requests lists from specific regions, clearly includes fewer
references of the most “popular” responses and is more likely to include long-tail responses.
Finally, to provide specific examples of the output for a set of identifiable entities, Figures 8 and
9 present the frequencies for the top 10 most popular entities and a few selected other examples
from the rest of the distribution. While Aristotle makes up as much as 20% of all entities identified
for Prompts 1 and 2, he is mentioned only half as much with versions 4 and 5. Correspondingly,
other responses such as Buddhism and Confucianism is more frequent with the prompts asking for
diversity, while more rare entities like Yoruba philosophy or Avicenna are almost never mentioned,
regardless of the prompt. The plots also underestimate the full list of 2,693 entities, as they are
truncated to the 600 most frequent responses.
5.4. Discussion
The results suggest that prompting strategy is one approach that can, to an extent, improve the
diversity of outputs. But in the corpus including all five prompts, Avicenna or Ibn Sadi appears 57
and 26 times respectively, for 83 total mentions, versus 1,080 mentions for Martha Nussbaum, and
al-Balkhı̄ al-Bukhārı̄ ” and is also written “Ibn Sina”,“Sharaf al-Mulk”, etc.
TABLE 3. Shannon and Pielou’s Indices by Model and Prompt
4,779 for Aristotle.16 The concept of ‘Ubuntu’, which has become relatively well-known among the
general population for a non-western philosophical notion, receives 126 mentions.
The limitations of the approach urge caution when making specific claims of cultural bias,
but we think that further investigation is merited to better understand not only problems of
cultural representativeness but of the general issue of ‘homogenization’ or bias towards popular or
common responses identified in the model. One caveat is that the specific wording of the prompts,
such as the phrases “well-being” and “philosopher”, might bias the results against inclusion of
more diverse ethical traditions grounded in religions or less associated with the specific phrase.
Furthermore, additional work could investigate the impact of alternative prompting approaches,
different temperature settings and other decoding strategies (Wiher, Meister, and Cotterell 2022).
This specific philosophic example is obviously limited as well, and could be extended to other
domains.
6. Conclusion
0.03
0.02
0.01
0.00
gpt-3.5-turbo
0.05
v1
v5
0.04
0.03
0.02
0.01
0.00
claude-3-sonnet
0.05
v1
v5
0.04
0.03
0.02
0.01
0.00
FIGURE 7. Frequency of philosophical entities based on prompts 1 vs. 5, where (1) simply asks
for a response, while (5) asks for a list of twenty individuals or schools of thought from a specific
named region. Note: the x-axis is truncated to the 600 most frequent responses, out of 2,693 entities
identified in the corpus.
Frequency Frequency Frequency
0.00
0.05
0.10
0.15
0.20
0.25
0.30
0.00
0.05
0.10
0.15
0.20
0.25
0.30
0.00
0.05
0.10
0.15
0.20
0.25
0.30
Ra Ra Ra
w ls w ls w ls
Zh Zh Zh
ua ua ua
ng ng n
zi z i
gz
i
Co Co Co
nfu nfu nfu
cia cia c ian
nis nis
m m ism
Ka Ka Ka
Ph n t Ph nt Ph n t
en en en
om om om
en e no en
olo log olo
gy y gy
Mo Mo Mo
his his his
m m m
Bu Bu Bu
dd dd dd
his his his
m m m
Prompt v3
Prompt v2
gemini-pro
gemini-pro
gemini-pro
llama2-70b
llama2-70b
llama2-70b
gpt-3.5-turbo
gpt-3.5-turbo
gpt-3.5-turbo
claude-3-sonnet
claude-3-sonnet
claude-3-sonnet
Frequency Frequency
0.00
0.05
0.10
0.15
0.20
0.25
0.30
0.00
0.05
0.10
0.15
0.20
0.25
0.30
Ari Ari
sto sto
Jer tle Jer tle
em em
yB yB
en en
tha tha
m m
Uti Uti
lita lita
ria ria
nis nis
Jea m Jea m
n-P n-P
au au
lS lS
art art
re re
Joh Joh
nS nS
tua tua
rt M rt M
ill ill
Ra Ra
w ls w ls
Zh Zh
ua ua
ng ng
zi z i
Co Co
nfu nfu
cia cia
nis nis
m m
Pla Pla
Ma to Ma to
rth rth
Sto Sto
ics ics
Ka Ka
Ph n t Ph nt
en en
om om
en e no
olo log
gy y
Mo Mo
his his
m m
Bu Bu
dd dd
his his
m m
Tao Tao
ism ism
Hin Hin
du du
ism ism
Jai Jai
nis nis
Ha m Ha m
us us
ap ap
hil hil
oso o sop
Av ph Av
ic en y ic en
hy
na na
(Ib ( Ibn
nS Sin
ina
) a)
Prompt v4
Prompt v5
Sik Sik
his his
Ub
un m Ub
un m
tu tu
p hil ph
oso ilo
ph sop
Yor
u y Yor
ub
hy
ba ap
ph hil
ilo oso
sop p
hy hy
Model
Model
gemini-pro
gemini-pro
gpt-3.5-turbo
gpt-3.5-turbo
claude-3-sonnet
claude-3-sonnet
FIGURE 9. Frequency for selected mentions. Because prompt v5 encourages greater diversity, there
References
Angelucci, C.; Cagé, J.; and Sinkinson, M. forthcoming. Media Competition and News Diets.
American Economic Journal: Microeconomics.
Anil, R.; Borgeaud, S.; Wu, Y.; Alayrac, J.-B.; Yu, J.; Soricut, R.; Schalkwyk, J.; Dai, A. M.; and Hauth,
A. 2023. Gemini: A family of highly capable multimodal models.
Arora, S.; Ge, R.; Liang, Y.; Ma, T.; and Zhang, Y. 2017. Generalization and equilibrium in generative
adversarial nets (gans). In International conference on machine learning, 224–232. PMLR.
Bakshy, E.; Rosenn, I.; Marlow, C.; and Adamic, L. 2012. The role of social networks in information
diffusion. In Proceedings of the 21st international conference on World Wide Web, WWW ’12, 519–528.
New York, NY, USA: Association for Computing Machinery.
Ball, Z., and Lewis, K. 2018. Mass Collaboration Project Recommendation Within Open-Innovation
Design Networks. Journal of Mechanical Design 141(021105).
Banerjee, A. V. 1992. A Simple Model of Herd Behavior. The Quarterly Journal of Economics
107(3):797–817.
Barberá, P. 2020. Social media, echo chambers, and political polarization. In Social Media and
Democracy: The State of the Field, Prospects for Reform. Cambridge University Press. 34–55.
Barbieri, N.; Bonchi, F.; and Manco, G. 2013. Topic-aware social influence propagation models.
Knowledge and Information Systems 37(3):555–584.
Barrat, A.; Barthélemy, M.; and Vespignani, A. 2008. Dynamical Processes on Complex Networks.
Cambridge University Press.
Bikhchandani, S.; Hirshleifer, D.; and Welch, I. 1998. Learning from the Behavior of Others:
Conformity, Fads, and Informational Cascades. The Journal of Economic Perspectives 12(3):151–170.
Publisher: American Economic Association.
Bohacek, M., and Farid, H. 2023. Nepotistically Trained Generative-AI Models Collapse.
arXiv:2311.12202 [cs].
Boone, C.; Carroll, G. R.; and van Witteloostuijn, A. 2002. Resource Distributions and Market
Partitioning: Dutch Daily Newspapers, 1968 TO 1994. American Sociological Review 67(3):408–431.
Brynjolfsson, E.; Hu, Y. J.; and Smith, M. D. 2003. Consumer Surplus in the Digital Economy:
Estimating the Value of Increased Product Variety at Online Booksellers. Management Science
49(11):1580–1596.
Brynjolfsson, E.; Hu, Y. J.; and Smith, M. D. 2006. From niches to riches: Anatomy of the long tail.
Sloan management review 47(4):67–71.
Cagé, J. 2016. Saving the media: Capitalism, crowdfunding, and democracy. Harvard University Press.
Cagé, J. 2020. Media competition, information provision and political participation: Evidence from
French local newspapers and elections, 1944–2014. Journal of Public Economics 185:104077.
Castellano, C.; Fortunato, S.; and Loreto, V. 2009. Statistical physics of social dynamics. Reviews of
modern physics 81(2):591.
Centola, D. 2010. The Spread of Behavior in an Online Social Network Experiment. Science
329(5996):1194–1197. Publisher: American Association for the Advancement of Science.
Chen, C., and Shu, K. 2023. Combating Misinformation in the Age of LLMs: Opportunities and
Challenges. arXiv:2311.05656 [cs].
Chen, L.; Chen, P.; and Lin, Z. 2020. Artificial Intelligence in Education: A Review. IEEE Access
8:75264–75278.
Chen, L.; Razniewski, S.; and Weikum, G. 2023. Knowledge Base Completion for Long-Tail Entities.
arXiv:2306.17472 [cs].
Cherniack, S. 1994. Book Culture and Textual Transmission in Sung China. Harvard Journal of
Asiatic Studies 54(1):5–125. Publisher: Harvard-Yenching Institute.
Chonka, P.; Diepeveen, S.; and Haile, Y. 2023. Algorithmic power and African indigenous languages:
search engine autocomplete and the global multilingual Internet. Media, Culture & Society
45(2):246–265. Publisher: SAGE Publications Ltd.
Christian, B. 2021. The alignment problem: How can machines learn human values? Atlantic Books.
Cinelli, M.; De Francisci Morales, G.; Galeazzi, A.; Quattrociocchi, W.; and Starnini, M. 2021.
The echo chamber effect on social media. Proceedings of the National Academy of Sciences
118(9):e2023301118.
Cinus, F.; Minici, M.; Monti, C.; and Bonchi, F. 2022. The Effect of People Recommenders on Echo
Chambers and Polarization. Proceedings of the International AAAI Conference on Web and Social
Media 16:90–101.
Das, D.; De Langis, K.; Martin-Boyle, A.; Kim, J.; Lee, M.; Kim, Z. M.; Hayati, S. A.; Owan, R.; Hu,
B.; Parkar, R.; Koo, R.; Park, J.; Tyagi, A.; Ferland, L.; Roy, S.; Liu, V.; and Kang, D. 2024. Under
the Surface: Tracking the Artifactuality of LLM-Generated Data. arXiv:2401.14698 [cs].
De Courson, B.; Fitouchi, L.; Bouchaud, J.-P.; and Benzaquen, M. 2021. Cultural diversity and
wisdom of crowds are mutually beneficial and evolutionarily stable. Scientific Reports 11(1):16566.
Delile, J.; Mukherjee, S.; Van Pamel, A.; and Zhukov, L. 2024. Graph-Based Retriever Captures the
Long Tail of Biomedical Knowledge. arXiv:2402.12352 [cs].
Dittmar, J. E. 2011. Information Technology and Economic Change: The Impact of The Printing
Press *. The Quarterly Journal of Economics 126(3):1133–1172.
Dohmatob, E.; Feng, Y.; Yang, P.; Charton, F.; and Kempe, J. 2024. A Tale of Tails: Model Collapse
as a Change of Scaling Laws. arXiv:2402.07043 [cs].
Douglas, S. J. 2002. Mass media: From 1945 to the present. In Agnew, J.-C., and Rosenzweig, R.,
eds., A Companion to Post-1945 America, volume 18. Wiley Online Library. 78–95.
Eisenstein, E. L. 1980. The Printing Press as an Agent of Change. Cambridge University Press.
Festinger, L.; Schachter, S.; and Back, K. 1950. Social pressures in informal groups; a study of human
factors in housing. Social pressures in informal groups; a study of human factors in housing.
Oxford, England: Harper. Pages: 240.
Fieser, J., and Dowden, B. 2024. Internet encyclopedia of philosophy. https://iep.utm.edu/. ISSN
2161-0002.
Fisher, L. 2024. UK government to trial ‘red box’ AI tools to improve ministerial efficiency.
Gao, C.; Wang, S.; Li, S.; Chen, J.; He, X.; Lei, W.; Li, B.; Zhang, Y.; and Jiang, P. 2023. CIRS:
Bursting Filter Bubbles by Counterfactual Interactive Recommender System. ACM Transactions
on Information Systems 42(1):14:1–14:27.
Gesi, J.; Shen, X.; Geng, Y.; Chen, Q.; and Ahmed, I. 2023. Leveraging Feature Bias for Scalable
Misprediction Explanation of Machine Learning Models. In 2023 IEEE/ACM 45th International
Conference on Software Engineering (ICSE), 1559–1570. ISSN: 1558-1225.
Goldenberg, J.; Libai, B.; and Muller, E. 2001. Talk of the Network: A Complex Systems Look at the
Underlying Process of Word-of-Mouth. Marketing Letters 12(3):211–223.
Goodfellow, I. 2016. Nips 2016 tutorial: Generative adversarial networks. arXiv preprint
arXiv:1701.00160.
Graham, R. 2023. The ethical dimensions of Google autocomplete. Big Data & Society
10(1):20539517231156518. Publisher: SAGE Publications Ltd.
Gruhl, D.; Guha, R.; Liben-Nowell, D.; and Tomkins, A. 2004. Information diffusion through
blogspace. In Proceedings of the 13th International Conference on World Wide Web, 491–501.
Guo, Y.; Shang, G.; Vazirgiannis, M.; and Clavel, C. 2023. The Curious Decline of Linguistic Diversity:
Training Language Models on Synthetic Text. arXiv:2311.09807 [cs].
Hackforth, R. 1972. Plato: Phaedrus. Cambridge University Press.
Havelock, E. A. 2019. The Literate Revolution in Greece and Its Cultural Consequences. Princeton
University Press.
Hegel, G. W. F. 2018. Hegel: The phenomenology of spirit. Oxford University Press.
Heidari, A.; Jafari Navimipour, N.; Dag, H.; and Unal, M. 2023. Deepfake detection using deep
learning methods: A systematic and comprehensive review. Wiley Interdisciplinary Reviews: Data
Mining and Knowledge Discovery 14:1520.
Henrich, J. 2004. Demography and cultural evolution: How adaptive cultural processes can produce
maladaptive losses—the tasmanian case. American Antiquity 69(2):197–214.
Herder, J. G. 2024. Ideas for the Philosophy of the History of Mankind. Princeton University Press.
Hoffmann, J.; Borgeaud, S.; Mensch, A.; Buchatskaya, E.; Cai, T.; Rutherford, E.; Casas, D. d. L.;
Hendricks, L. A.; Welbl, J.; and Clark, A. 2022. Training compute-optimal large language
models.
Holtzman, A.; Buys, J.; Du, L.; Forbes, M.; and Choi, Y. 2020. The Curious Case of Neural Text
Degeneration. arXiv:1904.09751 [cs].
Jamieson, K. H., and Cappella, J. N. 2008. Echo Chamber: Rush Limbaugh and the Conservative Media
Establishment. Oxford University Press.
Jiang, R.; Chiappa, S.; Lattimore, T.; György, A.; and Kohli, P. 2019. Degenerate Feedback Loops
in Recommender Systems. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and
Society, 383–390.
Kandpal, N.; Deng, H.; Roberts, A.; Wallace, E.; and Raffel, C. 2023. Large Language Models
Struggle to Learn Long-Tail Knowledge. In Proceedings of the 40th International Conference on
Machine Learning, 15696–15707. PMLR. ISSN: 2640-3498.
Kant, I. 1933. Critique of pure reason (Norman Kemp Smith, translator). New York: The Modem
Library.
Karlaš, B.; Dao, D.; Interlandi, M.; Li, B.; Schelter, S.; Wu, W.; and Zhang, C. 2022. Data Debugging
with Shapley Importance over End-to-End Machine Learning Pipelines. arXiv:2204.11131 [cs].
Keijzer, M. A., and Mäs, M. 2022. The complex link between filter bubbles and opinion polarization.
Data Science 5(2):139–166.
Kermack, W. O., and McKendrick, A. G. 1927. A contribution to the mathematical theory of
epidemics. Proceedings of the royal society of london. Series A, Containing papers of a mathematical
and physical character 115(772):700–721.
Khodadi, M.; Allahyari, A.; Vagnozzi, S.; and Mota, D. F. 2020. Black holes with scalar hair in light
of the event horizon telescope. Journal of Cosmology and Astroparticle Physics 2020(09):026.
Klug, D.; Qin, Y.; Evans, M.; and Kaufman, G. 2021. Trick and please. a mixed-method study on user
assumptions about the tiktok algorithm. In Proceedings of the 13th ACM Web Science Conference
2021, 84–92.
Kramár, J.; Lieberum, T.; Shah, R.; and Nanda, N. 2024. AtP*: An efficient and scalable method for
localizing LLM behaviour to components. arXiv:2403.00745 [cs].
Kuhn, T. S. 1997. The structure of scientific revolutions, volume 962. University of Chicago press
Chicago.
Kürnsteiner, P.; Wilms, M. B.; Weisheit, A.; Gault, B.; Jägle, E. A.; and Raabe, D. 2020. High-
strength Damascus steel by additive manufacturing. Nature 582(7813):515–519. Publisher:
Nature Publishing Group.
Layton, B. 1989. The significance of basilides in ancient christian thought. Representations 28:135–
151.
Li, H.; Ning, Y.; Liao, Z.; Wang, S.; Li, X. L.; Lu, X.; Brahman, F.; Zhao, W.; Choi, Y.; and Ren, X.
2023. In Search of the Long-Tail: Systematic Generation of Long-Tail Knowledge via Logical
Rule Guided Search. arXiv:2311.07237 [cs].
Lin, A.; Wang, J.; Zhu, Z.; and Caverlee, J. 2022. Quantifying and Mitigating Popularity Bias in
Conversational Recommender Systems. In Proceedings of the 31st ACM International Conference
on Information & Knowledge Management, CIKM ’22, 1238–1247. New York, NY, USA: Association
for Computing Machinery.
Mannheim, K. 1952. The sociological problem of generations. Essays on the Sociology of Knowledge
306:163–195.
Melis, G.; György, A.; and Blunsom, P. 2022. Mutual information constraints for Monte-Carlo
objectives to prevent posterior collapse especially in language modelling. The Journal of Machine
Learning Research 23(1):75:3266–75:3301.
Mercier, H., and Miton, H. 2019. Utilizing simple cues to informational dependency. Evolution and
Human Behavior 40(3):301–314.
Mesoudi, A., and Whiten, A. 2008. The multiple roles of cultural transmission experiments
in understanding human cultural evolution. Philosophical Transactions of the Royal Society B:
Biological Sciences 363(1509):3489. Publisher: The Royal Society.
Mokyr, J. 2011. The Gifts of Athena: Historical Origins of the Knowledge Economy. In The Gifts of
Athena. Princeton University Press.
Nazer, L. H.; Zatarah, R.; Waldrip, S.; Ke, J. X. C.; Moukheiber, M.; Khanna, A. K.; Hicklen, R. S.;
Moukheiber, L.; Moukheiber, D.; Ma, H.; and Mathur, P. 2023. Bias in artificial intelligence
algorithms and recommendations for mitigation. PLOS Digital Health 2(6):e0000278. Publisher:
Public Library of Science.
Nettle, D., and Romaine, S. 2000. Vanishing Voices: The Extinction of the World’s Languages. Oxford
University Press.
Nowak, A.; Szamrej, J.; and Latané, B. 1990. From private attitude to public opinion: A dynamic
theory of social impact. Psychological Review 97(3):362–376. Place: US Publisher: American
Psychological Association.
Ong, W. J. 2013. Orality and Literacy: 30th Anniversary Edition. Routledge.
Opdahl, A. L.; Tessem, B.; Dang-Nguyen, D.-T.; Motta, E.; Setty, V.; Throndsen, E.; Tverberg, A.; and
Trattner, C. 2023. Trustworthy journalism through ai. Data & Knowledge Engineering 146:102182.
O’Reilly, T. 2005. What Is Web 2.0.
Pariser, E. 2011. The filter bubble: What the Internet is hiding from you. penguin UK.
Pfister, D. S. 2011. The Logos of the Blogosphere: Flooding the Zone, Invention, and Attention in
the Lott Imbroglio. Argumentation and Advocacy 47(3):141–162.
Pielou, E. C. 1966. The measurement of diversity in different types of biological collections. Journal
of theoretical biology 13:131–144.
Russo, L., et al. 2003. The forgotten revolution: how science was born in 300 BC and why it had to be
reborn. Springer Science & Business Media.
Schubert, E.; Sander, J.; Ester, M.; Kriegel, H. P.; and Xu, X. 2017. Dbscan revisited, revisited: why
and how you should (still) use dbscan. ACM Transactions on Database Systems (TODS) 42(3):1–21.
Seymour, L. M.; Maragh, J.; Sabatini, P.; Di Tommaso, M.; Weaver, J. C.; and Masic, A. 2023. Hot
mixing: Mechanistic insights into the durability of ancient Roman concrete. Science Advances
9(1):eadd1602. Publisher: American Association for the Advancement of Science.
Sharma, N.; Liao, Q. V.; and Xiao, Z. 2024. Generative Echo Chamber? Effects of LLM-Powered
Search Systems on Diverse Information Seeking. arXiv:2402.05880 [cs].
Shumailov, I.; Shumaylov, Z.; Zhao, Y.; Gal, Y.; Papernot, N.; and Anderson, R. 2023. The curse of
recursion: Training on generated data makes models forget.
Smith, L., and Sørensen, P. 2000. Pathological Outcomes of Observational Learning. Econometrica
68(2):371–398. _eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1111/1468-0262.00113.
Su, Y.; Lan, T.; Wang, Y.; Yogatama, D.; Kong, L.; and Collier, N. 2022. A Contrastive Framework for
Neural Text Generation. In Koyejo, S.; Mohamed, S.; Agarwal, A.; Belgrave, D.; Cho, K.; and
Oh, A., eds., Advances in Neural Information Processing Systems, volume 35, 21548–21561. Curran
Associates, Inc.
Taleb, N. N. 2007. Black swans and the domains of statistics. The American Statistician 61(3):198–200.
Tversky, A., and Kahneman, D. 1973. Availability: A heuristic for judging frequency and probability.
Cognitive Psychology 5(2):207–232.
Weil, P. 2008. Overlapping Generations: The First Jubilee. Journal of Economic Perspectives 22(4):115–
134.
Wendler, C.; Veselovsky, V.; Monea, G.; and West, R. 2024. Do Llamas Work in English? On the
Latent Language of Multilingual Transformers. arXiv:2402.10588 [cs].
Wiher, G.; Meister, C.; and Cotterell, R. 2022. On Decoding Strategies for Neural Text Generators.
Transactions of the Association for Computational Linguistics 10:997–1012.
Wu, Z.; Geiger, A.; Icard, T.; Potts, C.; and Goodman, N. 2023. Interpretability at Scale: Identifying
Causal Mechanisms in Alpaca. Advances in Neural Information Processing Systems 36:78205–78226.
Wu, T. 2011. The master switch: The rise and fall of information empires. Vintage.
Yu, Y.; Zhuang, Y.; Zhang, J.; Meng, Y.; Ratner, A. J.; Krishna, R.; Shen, J.; and Zhang, C. 2023. Large
Language Model as Attributed Training Data Generator: A Tale of Diversity and Bias. Advances
in Neural Information Processing Systems 36:55734–55784.
Zamora Bonilla, J. P. 2006. Science Studies and the Theory of Games. Perspectives on Science
14(4):525–557.
Zamora-Bonilla, J. 2010. What Games Do Scientists Play? Rationality and Objectivity in a Game-
Theoretic Approach to the Social Construction of Scientific Knowledge. In Suárez, M.; Dorato,
M.; and Rédei, M., eds., EPSA Epistemology and Methodology of Science: Launch of the European
Philosophy of Science Association. Dordrecht: Springer Netherlands. 323–332.
Appendix A
As mentioned above, the reported results used a t-distribution with 10 degrees of freedom, which
has slightly wider tails than a standard normal distribution. We can compare the results with a
standard normal distribution (i.e.a t-distribution as the degrees of freedom becomes large) or
with wider tails. In Figure A1, we plot a comparison of the results from the main section (with 10
degrees of freedom with wider or narrower tails (3 and 9999 degrees of freedom respectively). The
main difference is for more extreme discounts provided by AI (< 0.7), for which the wider tails
contribute to knowledge collapse (i.e.generate a public knowledge distribution further from the
true distribution). Narrower tails, such as from a standard normal distribution, generate results
broadly similar to the main model. Thus, as expected more information in the tails makes the
effect of knowledge collapse more pronounced, but is plays less of a role than the other parameters
discussed above in determining the dynamic of collapse.
0.5
0.4
0.3
0.2
0.2 0.4 0.6 0.8 1.0
Discount factor
To define knowledge collapse we need to distinguish between a few conceptual sets of ‘knowledge’,
whether or not these are empirically observable.17 First, we consider the broad set of historical
17
The broadest definition of ‘human knowledge’ might encompass all the beliefs, information, values, and representa-
tions of the world ever held by humans anywhere on earth, whether recorded or not. We are unable to access almost all
human knowledge that was at one point held in common within communities of humans, shared
and reproduced in a regular way, which we might call ‘broad historical knowledge’.
Second, we consider the set of knowledge that is held or accessible to us, (humans who are
living in a given epoch), which we call ‘available current knowledge.’ In the example cited in the
main section, the ancient Roman recipe for concrete is part of broad historical knowledge but not
part of available current knowledge.
Technological innovations from the printing press to the internet to AI mediate human interac-
tions and human’s exposure to historical and current sources of knowledge. The net effect might
be to restrict or expand access to diverse knowledge and the long-tails of human knowledge. For
example, the digitization of archives might make obscure sources available to a wider audience
and thus increase the amount of ‘broad historical knowledge’ that is part of the ‘available current
knowledge.’
We also distinguish a third, narrower set of knowledge, which reflects not what is theoretically
accessible to humans but which is readily part of human patterns of thinking or habits of thought.
This we call ‘human memory knowledge’ or ‘human working knowledge’ by reference to human
working memory.
For example, consider the problem of listing all the animals that have ever existed on earth.
There might be some that humans previously knew about, but which subsequently went extinct
and which do not exist anywhere among the scientific literature or individuals currently living on
earth. More narrowly, the set of “available current knowledge” corresponds to the set of all animals
that a team of all biologists could compile with access to the internet and other records. Finally,
however, if we were able to conduct a survey of all humans on earth and ask them to name as many
animals as possible in, say, one day, we would come up with a more limited list (that would include
many repetitions).
In many practical applications, ‘human working knowledge’ is the most relevant because it is
the knowledge that shapes human action and reflection. A doctor considering possible sources
of a crossover pathogen might rely on their knowledge of common species in asking a patient if
they had recently been in the presence of certain animals (even if a researcher who specializes
in this area might consult know more and sources to find a longer possible list). A linguist trying
to evaluate or create possible linguistic theories implicitly bases their judgement on the known
language families and their structures, and so on. Edison and his team famously tried thousands of
different filament materials, but if it bamboo had not been among the materials that came to mind
of this, and we tend to assume that the useful parts of this have been passed on to others, but theoretically we might
want to allow for the fact that, for example some human somewhere once had an important, original, and useful belief
just before they, say, got hit by a car and could not tell anyone. Secondly, in using the term ‘knowledge’, we do not restrict
our focus based on the truth of the beliefs held, such that in referring to ‘human knowledge’ we refer to a variety of
beliefs and statements, some of which contract others.
as they searched alternatives, a practical electric bulb may have been invented only later.
Finally, it is useful to define the ‘epistemic horizon’ as the set of knowledge that a community
of humans considers practically possible to know and worth knowing.18 A common controversy in
the public imagination is whether traditional medicines are worth consideration when searching
for medical cures. Such traditional medicines might be outside of the epistemic horizon because
they are not written down in the scientific literature, are only known by individuals speaking lesser
known languages, or because the scientists in question consider them too costly to acquire or
unlikely to be beneficial. One way to think about this relationship is as a generalization of ‘availability
bias’, in which we take the set of readily recalled information to be more likely, important, or relevant
(Tversky and Kahneman 1973).
In these terms, we define ‘knowledge collapse’ as the progressive narrowing over time (or over
technological representations) of the set of human working knowledge and the current human
epistemic horizon relative to the set of broad historical knowledge.
DEFINITION A1 (KNOWLEDGE COLLAPSE). Consider a historical set of human knowledge ptrue (x)
represented as a distribution over time or technological representations. We define knowledge collapse as
the progressive narrowing of this distribution such that the variance of ptrue (x)t , denoted as σ2 (ptrue (x)t ),
approaches zero as time t progresses. Formally, we can represent this as:
In the limit, the probability density function of ptrue (x)t approaches a Dirac delta function centered at
some knowledge point µt , i.e.,
d
(A2) ptrue (x)t Ð
→ δ(µt )
d
where Ð
→ denotes convergence in distribution.
On a theoretical level, the idea of epistemic horizon has an intellectual heritage in Immanuel
Kant’s argument about the forms and categories of understanding that underly the possibility of
knowledge (Kant 1933). Subsequent authors expanded on the implications if these categories are in
some way fashioned by one’s upbringing and community (Hegel 2018; Mannheim 1952).19 A related
concern is the way that the scientific community can be, at least during certain epochs, bounded
18
In economic terms, it is the set of information that for which the individual believes the expected returns are greater
than the expected costs. This might be considered for a specific task or set of tasks, but could be generalized to the set of
knowledge for which she expects positive gains over a period of time, her lifetime, or for society over a finite or infinite
horizon with discounting.
19
e.g.“If man received every thing from himself and developed it independently of extrinsic objects, then a history of
a man might be possible, but not of men in general. But as our specific character resides precisely in this, that, born
almost without instinct, we are raised to manhood only by lifelong practice, on which both the perfectibility as well as
by its inherited understanding of the world (Kuhn 1997; Zamora Bonilla 2006). As noted above,
specific technological forms may generate a flood of information that inhibit the communication
of information (Pfister 2011).
Finally, one of the challenges presented by the ‘epistemic horizon’ (as of that of an ‘event
horizon’20 ) is that we cannot observe directly its limits. For example, the presence of an event
our current model takes to be very rare (e.g.a “20-sigma” event) can suggest our current model is
incorrect, but in the absence of such a rare event, we cannot know if the current tails of knowledge
are correct or too thin (Taleb 2007). These considerations suggest the concern of generational
knowledge collapse is plausible and an unbounded optimism in the ability of rational actors to
update on the value of tail knowledge may be shortsighted.
the corruptibility of our species rests, so it is precisely thereby that the history of mankind is made a whole: that is, a
chain of sociability and formative tradition from the first link to the last.(Herder 2024)
20
Technically, we can observe the ‘shadow’ of an event horizon (Khodadi et al. 2020)