Large language models collaborating on long-context tasks
A simple but effective approach to improve long-context understanding Previous studies have mainly explored two major directions: input reduction and window extension. Input reduction reduces the length of the input context — for example, by directly truncating the input — before feeding to downstream LLMs. RAG extends this direction by breaking the input into chunks […]