AI Navigate

Large Language Models and Book Summarization: Reading or Remembering, Which Is Better?

arXiv cs.CL / 3/12/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Large-context LLMs with windows reaching millions of tokens can process entire books in a single prompt.
  • For well-known books, LLMs can generate summaries based on internal knowledge acquired during training without reading the full text.
  • The study experimentally compares summaries produced from internal memory and from the full text of the book.
  • In general, having the full text yields more detailed summaries, but for some books internal-knowledge summaries perform better.
  • These results raise questions about long-text summarization capabilities, since information learned during training can outperform summarizing the full text in certain cases.

Abstract

Summarization is a core task in Natural Language Processing (NLP). Recent advances in Large Language Models (LLMs) and the introduction of large context windows reaching millions of tokens make it possible to process entire books in a single prompt. At the same time, for well-known books, LLMs can generate summaries based only on internal knowledge acquired during training. This raises several important questions: How do summaries generated from internal memory compare to those derived from the full text? Does prior knowledge influence summaries even when the model is given the book as input? In this work, we conduct an experimental evaluation of book summarization with state-of-the-art LLMs. We compare summaries of well-known books produced using (i) only the internal knowledge of the model and (ii) the full text of the book. The results show that having the full text provides more detailed summaries in general, but some books have better scores for the internal knowledge summaries. This puts into question the capabilities of models to perform summarization of long texts, as information learned during training can outperform summarization of the full text in some cases.