Are LLMs Prescient? A Continuous Evaluation using Daily News as the Oracle

New York University

Abstract

Many existing evaluation benchmarks for Large Language Models (LLMs) quickly become outdated due to the emergence of new models and training data. These benchmarks also fall short in assessing how LLM performance changes over time, as they consist of static questions without a temporal dimension. To address these limitations, we propose using future event prediction as a continuous evaluation method to assess LLMs' temporal generalization and forecasting abilities. Our benchmark, Daily Oracle, automatically generates question-answer (QA) pairs from daily news, challenging LLMs to predict "future" event outcomes. Our findings reveal that as pre-training data becomes outdated, LLM performance degrades over time. While Retrieval Augmented Generation (RAG) has the potential to enhance prediction accuracy, the performance degradation pattern persists, highlighting the need for continuous model updates.


Daily Oracle Dataset

Dataset Overview

Pie Category
Question Size

Example QA pairs

MY ALT TEXT

QA Construction Pipeline

For each day, we collect news articles from the daily-updated Common Crawl News Datase, and use LLM to generate QA pairs with the few-shot prompting technique.

MY ALT TEXT


Evaluation

Closed-Book Setting

MY ALT TEXT


Constrained Open-Book Setting

  • In the constrained open-book setting, we explore how access to news articles up to different time cutoffs influences LLM performance using RAG.
  • RAG cutoff: the latest accessible date for retrieving articles.
  • RAG has the potential to enhance prediction accuracy, the performance degradation pattern persists, highlighting the need for continuous model updates.


Gold Article Setting

MY ALT TEXT