LLM Context Windows

The cluster discusses limitations and capabilities of context windows in large language models (LLMs), including sizes like 8k-1M tokens, performance degradation with long contexts, comparisons between models, and strategies like summarization, RAG, or larger windows to handle them effectively.

➡️ Stable 1.6x AI & Machine Learning
4,343
Comments
15
Years Active
5
Top Authors
#9534
Topic ID

Activity Over Time

2009
1
2011
1
2012
1
2015
4
2016
2
2017
3
2018
3
2019
8
2020
35
2021
6
2022
61
2023
1,214
2024
944
2025
1,910
2026
154

Keywords

A1 C1 LLM RAG FWIW context.md SPR GPT API CAN context window llms tokens llm models model conversation larger contexts

Sample Comments

derwiki Jun 27, 2023 View on HN

Does it do well? How does it get around the context window limits?

randomjoe2 May 28, 2025 View on HN

Lots of models CAN handle large contexts, gemini 2.5 pro their latest model can take 1 million tokens of context

inciampati Nov 1, 2023 View on HN

Context is 8k and it's quadratic. It "sees" everything in that window. If you want to have a long conversation try Claude or some of the 32k models. Claude uses a strange kind of attention that isn't always as precise but it's very good at finding key information in huge documents.

brulard Aug 13, 2025 View on HN

I think you misunderstand how context in current LLMs works. To get the best results you have to be very careful to provide what is needed for immediate task progression, and postpone context thats needed later in the process. If you give all the context at once, you will likely get quite degraded output quality. Thats like if you want to give a junior developer his first task, you likely won't teach him every corner of your app. You would give him context he needs. It is similar with these

Alifatisk Jan 20, 2025 View on HN

This is impressive, how do people handle the limited context window of 64k tokens?

ginger_beer_m Apr 25, 2025 View on HN

Tiny context window is killing it.

nbsande Oct 29, 2023 View on HN

The large(100k tokens) context window together with the fact that it can actually use the information in that context window. From personal experience other models including open ai fail to properly answer when provided large(more than 5k tokens) inputs as context even when the model officially accepts much larger contexts. But Claude 2 models are uncannily good at taking all that context into consideration.

0x000xca0xfe Nov 3, 2023 View on HN

Are there any drawbacks to the larger context window? Like more hallucinations or lower speed?

k5hp May 12, 2023 View on HN

That's interesting indeed. Could be that the context window of the model is too small...

ptdn Jan 21, 2024 View on HN

The context windows of LLMs are now significantly larger than 2048 tokens, and there are clever ways to autopopulate context window to remind it of things.