Estimating the Self-Consistency of LLMs

Robert Nowak

公開日: 2025/9/23

Abstract

Systems often repeat the same prompt to large language models (LLMs) and aggregate responses to improve reliability. This short note analyzes an estimator of the self-consistency of LLMs and the tradeoffs it induces under a fixed compute budget $B=mn$, where $m$ is the number of prompts sampled from the task distribution and $n$ is the number of repeated LLM calls per prompt; the resulting analysis favors a rough split $m,n\propto\sqrt{B}$.

Estimating the Self-Consistency of LLMs | SummarXiv | SummarXiv