The Half-Life of History

A personal experiment in measuring what humanity has forgotten

I watched David Reich's Big Think interview on ancient DNA1 and it rewired how I think about the past. Reich is a geneticist at Harvard who showed that entire populations left no written record at all. The Yamnaya replaced roughly 75% of Central European genetic ancestry around 4,500 years ago2. We only know because of DNA. Ghost populations existed, thrived, and vanished, detectable only as statistical traces in modern genomes3.

That raised a question I couldn't shake. If genetic signal decays exponentially with time (recombination chops ancestral blocks smaller each generation, drift erases rare lineages), does everything else decay the same way? Academic papers, biographical records, surviving sources, how often we mention a century in books. I wanted to measure it. I built this project using Claude Code, Anthropic's AI coding assistant, to collect the data, run the statistical analysis, and render the charts you see below.

The central finding: historical knowledge follows an exponential decay curve with a half-life of roughly 84 years (uncertainty range: 23 to 101). For every 84 years further back you look, the composite knowledge index approximately halves.

Five Thousand Years of Data, at a Glance

I assembled five distinct measures to capture the breadth of historical knowledge. Academic works (via OpenAlex7) counts research papers tagged to each period. Named individuals tallies Wikipedia biographical articles for people born in each century. Source proxy estimates surviving primary sources: cuneiform tablets and inscriptions for antiquity, manuscript production for the medieval period (following Buringh & Van Zanden 20094), and print title counts for the early modern and modern eras. Ngram discourse measures how frequently each century is mentioned in the Google Books English corpus6. And films, a supplementary cultural metric, counts movies set in each period.

The chart below plots all five metrics for every time period, on a logarithmic scale (use the buttons to switch to linear). Toggle individual metrics on or off by clicking the legend. The pattern is unmistakable: every single metric climbs steeply as you approach the present.

How to read this chart: Each line tracks a different type of historical evidence across 50 centuries. The vertical axis uses a logarithmic scale: each gridline represents a tenfold increase. Click any label in the legend to show or hide that metric. Use the Log/Linear buttons to switch scales.

The 20th century dominates every measure: 1.37 million Wikipedia biographies, 156,000 academic works, and an estimated 85 million publications. The 3rd millennium BC, by contrast, has just 385 known individuals and 50,000 surviving inscriptions. Recent history doesn't just have more. It has orders of magnitude more.

Knowledge Decays Like a Radioactive Isotope

Samuel Arbesman, in The Half-Life of Facts5, proposed that knowledge has a measurable decay rate. I tested this directly. After combining four metrics into a single composite index (using z-scores to put them on a common scale), I tested four candidate models: exponential decay, power law, logarithmic, and linear. The exponential model won decisively.

Just as uranium-238 decays at a predictable rate, historical knowledge follows a smooth exponential curve. The best-fit model gives a half-life of about 84 years. That means for every 84 years further back you look, the composite knowledge index roughly halves. The shaded band below shows the uncertainty range: the true half-life likely falls between 23 and 101 years.

The residual plot underneath reveals which centuries defy the trend. Points far from the zero line are centuries with unusual stories to tell. The 1st century CE spikes upward (driven by its biblical and Roman cultural footprint), while the 19th century overshoots the model's prediction.

When I went looking for prior work, I found that someone had already measured a version of the same thing. Michel et al., in their 2011 Science paper that launched Google Ngrams as a research tool6, tracked the fame trajectories of individual people by measuring how quickly a person's name rises and falls in book mentions. For people born around 1865, the post-peak half-life of fame was roughly 73 years. I arrived at a nearby number (~84 years) using an entirely different method: a composite of four metrics, z-scored, spanning five millennia instead of two centuries, measuring eras instead of individuals. A skeptic would note that Google Ngrams appears in both studies, and that is fair. But in my composite, Ngrams is one of four equally weighted metrics. The other three (academic papers, Wikipedia biographies, and surviving sources) have nothing to do with book mentions. Two independent approaches, different data, different scales, and yet half-lives within 15% of each other. That convergence suggests the number reflects something real about how collective memory works.

There may be a reason the number is 84 and not 200 or 500. The Egyptologist Jan Assmann drew a distinction between what he called communicative memory (what gets transmitted orally across three or four generations, roughly 80 to 100 years) and cultural memory, the kind that survives only when institutions write it down and teach it8. The 84-year half-life sits right at that boundary. It is approximately the point where the last person who heard the story firsthand dies, and whatever was not committed to text, stone, or institutional habit begins to vanish.

How to read this chart: Each blue dot is one century's combined knowledge score (an average of all four metrics, standardized to the same scale). The red curve is the best-fit exponential model. The shaded band shows statistical uncertainty. Below, the residual plot shows which centuries deviate most from the overall trend.

Historians Amplify the Decay, Not Compensate for It

The obvious objection is that of course we know more about recent history. More sources survived. The more interesting question is whether historians compensate for lost sources by studying ancient periods more intensely, or whether they make the imbalance worse.

The chart below answers this by comparing two curves: the green line tracks source availability (surviving tablets, manuscripts, and books), while the blue line tracks academic attention (OpenAlex publication counts). Where academic attention exceeds source availability, the green shading marks an era as overstudied. Where attention falls below sources, the red shading marks it as understudied.

How to read this chart: The green line tracks how many primary sources survive for each era. The blue line tracks how much academic attention that era receives. Green shading marks periods where scholars publish more than the surviving sources would predict (overstudied). Red shading marks periods where sources exceed scholarly attention (understudied).

The pattern is stark. Before the 12th century, every era is understudied. Historians give these periods less attention than their surviving source base would predict. From the 13th century onward, academic attention not only catches up but accelerates far beyond what sources alone would warrant. The 19th century is the most overstudied era, with a bias residual of +3.1, likely reflecting Anglophone historiographic priorities. The 21st century swings the other direction (3.5), but this is expected: the present has not yet become history.

One important caveat: the source proxy changes measurement method across eras (inscriptions in antiquity, manuscripts in the medieval period, print titles after Gutenberg). Cross-era comparisons are therefore approximate, not exact. But the overall direction of the bias is clear regardless.

What Drove the Acceleration

When all four metrics (plus the supplementary films count) are overlaid on a single timeline, the overall trajectory is a smooth upward climb, consistent with the exponential model from the previous chapter. But three historical transitions contributed disproportionately to that acceleration. Around the 1450s, the printing press began a slow transformation of source availability; the named-individuals and academic-works lines begin to steepen in the centuries after Gutenberg, though the effect takes generations to appear. Around 1800, mass literacy, state record-keeping, and industrialization pushed most metrics into steeper growth: the 19th century shows the largest single-century jumps across multiple measures. And the 20th century's digital infrastructure produced a final surge in some metrics, though not all: academic works barely grew from the 19th century, while named individuals and source volumes exploded.

Use the Log/Linear buttons below to switch between scales. On a linear scale, the 20th century towers so far above everything else that ancient history becomes a flat line at zero. The log scale reveals the continuous nature of the acceleration: there is no single moment where everything changes at once, but rather a compounding process that these three transitions each amplified.

How to read this chart: All five metrics plotted together on one timeline. Dashed vertical lines mark three historical transitions. Use the Log/Linear buttons to switch scales: linear makes the 20th century dominate; logarithmic reveals the full shape of the curve.

A Bird's-Eye View Reveals Dark Zones and Bright Hotspots

The heatmap below offers a compact view of all five metrics across all 33 time periods, with warmer colors indicating higher relative knowledge density (ranked within each metric, so colors reflect relative standing rather than absolute values). The dark zones (the 10th9th centuries BC, the early medieval period) are the deep troughs of human knowledge, eras where very few named individuals survive, minimal inscriptions exist, and academic attention is nearly absent.

In contrast, bright hotspots mark periods with disproportionate cultural footprints: Classical Greece (5th century BC), Imperial Rome (1st2nd century CE), and of course the entire modern era. The rightmost columns glow with intensity.

How to read this chart: Each cell is colored by rank within its row. Warmer colors mean higher relative scores; dark zones are periods where that metric is lowest. Read across a row to see how one metric varies over time. Read down a column to compare all metrics for one period.

One visible anomaly: the 1st century CE has an inflated ngram score. This is a data quality issue. The phrase first century appears frequently in non-historical contexts (the first century of the Industrial Revolution, the first century of American democracy), artificially boosting the signal. I'm flagging this rather than smoothing it away.

All Four Metrics Independently Agree

If academic papers, Wikipedia articles, surviving sources, and book mentions all point in the same direction, the overall pattern is robust. The scatter plots below show this directly. Each dot is one time period. When two metrics track each other, the dots line up along a diagonal. They do, consistently: Spearman rank correlations range from 0.78 to 0.93, all statistically significant after Benjamini-Hochberg correction.

This validates combining the metrics into a single composite index. They aren't telling four different stories. They are four witnesses to the same underlying phenomenon: the exponential erosion of historical knowledge over time.

Other researchers have approached the question with narrower scope. Candia et al. tracked how quickly specific cultural products (songs, movies, biographies) fade from online attention, finding a two-phase curve: a fast-decaying communicative component with a half-life of 20 to 30 years for biographies, and a slow-decaying cultural component that persists for centuries9. When Kestemont et al. applied mark-recapture methods from ecology to medieval manuscripts, they estimated that more than 90% of chivalric and heroic narratives from the period were lost entirely10. My composite index blends both layers of memory into a single curve. The 84-year half-life likely represents the handoff point between these two regimes of forgetting.

How to read this chart: Each dot is one time period. When two metrics agree (both high or both low for the same periods), dots line up along a diagonal. The r value measures this agreement on a scale from 0 (no relationship) to 1 (perfect agreement).

What This Analysis Cannot Tell Us

I cannot distinguish knowledge decay from evidence destruction. Fewer named individuals from the 7th century BC does not necessarily mean less was known or done in that era; it may mean fewer records survived, fewer were translated, or fewer were deemed notable by later generations. Survivorship bias runs through every metric here. The exponential curve may describe the decay of evidence as much as the decay of knowledge itself10.

The source proxy stitches together three different measurement regimes: inscription and tablet counts for antiquity, manuscript production estimates for the medieval period, and print title counts for the early modern and modern eras. Each transition (clay to parchment, parchment to print, print to digital) introduces a discontinuity. I have normalized within eras where possible, but cross-era comparisons remain approximate.

Of the six metrics originally collected, two turned out to be redundant or empty. The Wikipedia people count (wikipedia_people_count) was identical to named individuals for all 33 periods, both drawing from the same PetScan category query. Wikipedia event counts were null across the board (the expected category structure does not exist). After removing these, the analysis uses four distinct metrics, not five. This reduces the apparent independence of the composite index, though Cronbach's alpha (0.88) and the first principal component (explaining 75% of variance) suggest the remaining four metrics do measure a coherent underlying construct.

The confidence interval is wide: 23 to 101 years. The point estimate of ~84 years falls within the range reported by Michel et al. for cultural fame decay, and near the boundary of Assmann's communicative memory horizon. But with 32 data points and a composite built from heterogeneous sources, the precision of this estimate should not be overstated.

Finally, the exponential model is the best available parametric approximation among the four candidates tested, but the residuals show systematic structure. The Durbin-Watson statistic (0.17) indicates strong positive autocorrelation, and the Shapiro-Wilk test rejects normality. The model consistently overpredicts for ancient periods and underpredicts for the most recent centuries. This suggests the true relationship may involve regime shifts or a more complex functional form that a simple two-parameter model cannot capture.

What Gets Lost

The past is not equally preserved. It fades in a measurable, predictable way. An exponential curve as clean as any in physics. And the institutions we build to study history don't fight this decay. They amplify it, pouring attention onto the centuries that already have the most sources while the deep past grows quieter.

Reich showed that DNA preserves what texts cannot. But even DNA decays. Recombination chops ancestral blocks smaller each generation. Drift erases rare lineages. Everything in the historical record follows the same pattern. The signal fades. That's the finding.

This analysis reflects Anglophone and Western historiographic traditions. Chinese dynastic records, Islamic manuscript traditions, and Indian literary heritage would trace different curves. The half-life of history isn't a universal constant. It's a measure of what one civilization chose to preserve and what it chose to study.

Methodology & Technical Details

Model Selection

Four candidate models were fitted to the composite index (arithmetic mean of z-scores across four metrics, with wikipedia_people_count and wikipedia_events_count excluded) for 32 time periods (the 21st century was excluded as an incomplete period):

Exponential: AICc = 69.59, R² = 0.831Power law: AICc = 58.79, R² = 0.764Logarithmic: AICc = 48.31, R² = 0.672Linear: AICc = 22.53, R² = 0.266

The exponential model (y = a · eλt) fits best with parameters a = 6.69, λ = 0.00825. The AICc advantage over the power law (ΔAICc = 10.8) constitutes strong evidence. The linear null hypothesis (constant rate of decline) explains only 27% of variance and is decisively rejected.

Bootstrap Confidence Intervals

10,000 bootstrap resamples of the 32 data points yielded 95% confidence intervals for the decay parameter: λ = 0.00690.030 (median 0.0098). This translates to a half-life range of approximately 23101 years, with a point estimate of ~84 years.

Half-life (point): ~84 yearsHalf-life (95% CI): 23101 yearsBootstrap CI (λ): 0.00690.030N periods: 33 (32 fitted)

Composite Index Construction

Four metrics were used in the composite index: OpenAlex academic work counts, Wikipedia named individuals, era-specific source proxy, and Google Books ngram discourse frequency. Two metrics were excluded: wikipedia_people_count (identical to named individuals for all 33 periods, both drawing from the same PetScan query) and wikipedia_events_count (null for all periods). Films set in each period are collected and displayed but treated as supplementary and not included in the composite. Each metric was z-scored using century-only statistics, and the arithmetic mean of available z-scores per period formed the composite index.

Bias Residual

The bias chart compares z-scored source availability against z-scored academic attention (OpenAlex works). The residual (attention minus source) indicates whether an era receives more or less scholarly focus than its surviving source base would predict. The source proxy changes measurement regime across eras: cuneiform tablets and inscriptions for antiquity, manuscript production estimates (Buringh & Van Zanden 2009) for the medieval period, print title counts (USTC, Britannica) for the early modern era, and publication volume estimates for the modern period.

Correlation Analysis

Pairwise Spearman rank correlations were computed between the four metrics used in the composite. P-values were adjusted using the Benjamini-Hochberg FDR correction at α = 0.05. All pairs were statistically significant, with Spearman r values ranging from 0.78 to 0.93.

Data Sources

  • OpenAlex API (academic works, History concept C95457728)
  • PetScan / Wikipedia API (biography article counts)
  • Google Books Ngram Viewer (discourse frequency, en-2019 corpus)
  • Buringh & Van Zanden 2009 via Our World in Data (manuscript production)
  • USTC / ISTC (print edition counts)
  • Britannica / UNESCO (modern publication volumes)
  • EDH / PHI / CDLI (inscription/tablet databases)
  • Wikipedia Films categories (supplementary)

Sensitivity Analysis

To assess the stability of the half-life estimate, four sensitivity checks were performed:

Leave-one-metric-out: Dropping each of the four metrics in turn and refitting the exponential model yields half-lives ranging from 64 to 106 years. The estimate is most sensitive to dropping named individuals (half-life rises to 106 years) and least sensitive to dropping ngram discourse (64 years).

Drop openalex: 69 yrDrop named_individuals: 106 yrDrop source_proxy: 96 yrDrop ngram_discourse: 64 yr

Exclude millennia: Removing the two aggregated millennia (which span 10x the time of century-level data) and refitting yields a half-life of 84 years (N=30), virtually identical to the full estimate.

Exclude 1st century CE: Removing the 1st century (whose ngram value is inflated by non-historical usage) yields 84 years (N=31), again unchanged.

Median composite: Using the median instead of the mean to aggregate z-scores yields a half-life of 90 years, within the confidence interval.

Residual Diagnostics

Diagnostic tests on the exponential model residuals reveal systematic structure:

Durbin-Watson: 0.175Shapiro-Wilk p: 4.4 × 105Goldfeld-Quandt p: 1.00

The low Durbin-Watson statistic (0.175, where 2.0 indicates no autocorrelation) reveals strong positive autocorrelation in the residuals: the model systematically over- or under-predicts for consecutive time periods rather than scattering randomly. The Shapiro-Wilk test rejects normality (p < 0.001). There is no evidence of heteroscedasticity (Goldfeld-Quandt p = 1.00). These diagnostics indicate that while the exponential is the best of the four models tested, it does not fully capture the data-generating process.

Scale Reliability

To assess whether the four metrics measure a coherent underlying construct:

Cronbach's α: 0.879PCA PC1 variance: 74.8%

Cronbach's alpha of 0.879 exceeds the conventional threshold of 0.7 for acceptable internal consistency. The first principal component explains 74.8% of variance across the four metrics, confirming that a single latent dimension (plausibly historical knowledge density) accounts for most of the shared variation.

Model Comparison

Four candidate models were compared using AICc and R-squared:

Exponential: AICc = 69.59, R² = 0.831Power law: AICc = 58.79, R² = 0.764Logarithmic: AICc = 48.31, R² = 0.672Linear: AICc = 22.53, R² = 0.266

The exponential model fits best by both AICc (ΔAICc = 10.8 over the next-best power law) and R-squared. The linear null hypothesis (constant rate of decline) explains only 27% of variance and is decisively rejected.

Cross-Validation

Leave-one-out cross-validation RMSE for each model:

Exponential: RMSE = 0.663Logarithmic: RMSE = 0.560Linear: RMSE = 0.734Power law: RMSE = 5.800

The logarithmic model achieves the lowest LOOCV RMSE (0.560), slightly below the exponential (0.663). This suggests that the logarithmic model generalizes marginally better on a point-by-point basis, despite the exponential model's lower AICc. The discrepancy reflects the exponential model's sensitivity to extreme values (the 20th century), which dominate AICc but are downweighted in LOOCV. The power law's extremely high RMSE (5.800) indicates severe overfitting.

References

  1. Reich, D. (2026). Your ancestors aren't who you think they are [Interview]. Big Think. https://www.youtube.com/watch?v=a0uKLW07Jlg
  2. Haak, W., et al. (2015). Massive migration from the steppe was a source for Indo-European languages in Europe. Nature, 522, 207211.
  3. Lazaridis, I., et al. (2014). Ancient human genomes suggest three ancestral populations for present-day Europeans. Nature, 513, 409413.
  4. Buringh, E. & Van Zanden, J.L. (2009). Charting the Rise of the West: Manuscripts and Printed Books in Europe, A Long-Term Perspective. Journal of Economic History, 69(2), 409445.
  5. Arbesman, S. (2012). The Half-Life of Facts: Why Everything We Know Has an Expiration Date. Current/Penguin.
  6. Michel, J.B., et al. (2011). Quantitative Analysis of Culture Using Millions of Digitized Books. Science, 331, 176182.
  7. Priem, J., Piwowar, H. & Orr, R. (2022). OpenAlex: A fully-open index of scholarly works, authors, venues, institutions, and concepts. arXiv:2205.01833.
  8. Assmann, J. (1995). Collective Memory and Cultural Identity. New German Critique, 65, 125133.
  9. Candia, C., et al. (2019). The universal decay of collective memory and attention. Nature Human Behaviour, 3, 8291.
  10. Kestemont, M., et al. (2022). Forgotten books: The application of unseen species models to the survival of culture. Science, 375, 765769.
By Mark Pavlyukovskyy. Built with Claude Code. Data from OpenAlex, Wikipedia/PetScan, Google Ngrams, Our World in Data, USTC, EDH, PHI, CDLI. April 2026.