Positional Biases Shift as Inputs Approach Context Window Limits
Blerta Veseli¹, Julian Chibane², Mariya Toneva³, Alexander Koller¹Saarland Informatics Campus, Saarland University, Saarbrücken¹, Max Planck Institute for Informatics², Max Planck Institute for Software Systems³
COLM 2025, Montreal
Abstract
Large Language Models (LLMs) often struggle to use information across long inputs. Prior work has identified positional biases like the Lost in the Middle (LiM) effect—better performance when key information appears at the beginning (primacy) or end (recency) rather than in the middle—but long-context studies have not consistently replicated these effects. We address this by analyzing input length relative to each model’s context window and find that LiM is strongest when inputs occupy up to 50% of that window. Beyond that point, primacy weakens while recency remains relatively stable, effectively eliminating LiM and revealing a distance-based bias where performance is better when relevant information is closer to the end. Our results also suggest that successful retrieval is a prerequisite for reasoning, and that positional biases in reasoning are largely inherited from retrieval, with implications for long-context tasks, benchmark design, and evaluation.
Contributions
-
Defining relative input lengths.
Analysis of positional biases using relative input length (proportion of a model’s context window) rather than absolute lengths across models.
-
Positional biases are consistent across models when analyzed relative to a model's context window size.
-
LiM is strongest in input lengths up to 50% of a models context window size.
Brief interpretation of the figure.
-
LiM is strongest in input lengths up to 50% of a models context window size.
Brief interpretation of the figure.
-
With growing input size, primacy bias increasingly collapses, while the recency bias remains stable. This results in a more distance-based bias, i.e. accuracy higher when evidence is nearer the end.
Brief interpretation of the figure.
Acknowledgements
Funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) -- GRK 2853/1 “Neuroexplicit Models of Language, Vision, and Action” - project number 471607914.
Citation
@inproceedings{veseli2025positional,
title={Positional Biases Shift as Inputs Approach Context Window Limits},
author={Veseli, Blerta and Chibane, Julian and Toneva, Mariya and Koller, Alexander},
booktitle={Proceedings of the Conference on Language Modeling (COLM)},
year={2025}
}