Does very large context significantly increase a response time? Are there any benchmarks/leader-boards estimating different models in that regard?