▲ | brokegrammer 3 days ago | |||||||||||||
Many people are confused about the usefulness of 1M tokens because LLMs often start to get confused after about 100k. But this is big for Claude 4 because it uses automatic RAG when the context becomes large. With optimized retrieval thanks to RAG, we'll be able to make good use of those 1M tokens. | ||||||||||||||
▲ | m4r71n 3 days ago | parent [-] | |||||||||||||
How does this work under the hood? Does it build an in-memory vector database of the input sources and runs queries on top of that data to supplement the context window? | ||||||||||||||
|