| ▲ | ranger_danger 8 hours ago | |
where it makes sense IMO is when you need it to know about a large amount of information that's not already in the model, such as a company knowledgebase, code repositories or a trove of specialized legal documents... in that case it's not realistic to try to stuff the context window every time with that information, especially if you're trying to make a responsive chat bot. | ||
| ▲ | antirez 8 hours ago | parent | next [-] | |
With the current context windows and the ability those models did RL to work as agents, it's much faster and reliable for them to use tools and find the information before replying. Much better, no hallucinations problems (or a lot less), no fine tuning needed when information changes. I believe it is exactly in this case that fine tuning is no longer useful, and even in the past worked at very different degrees of quality. | ||
| ▲ | dotancohen 8 hours ago | parent | prev | next [-] | |
Wouldn't a RAG make more sense for this use case? | ||
| ▲ | larodi 5 hours ago | parent | prev [-] | |
indeed, and in practical terms, this is more often than never, and particularly with large knowledge bases. also makes super sense for VLMs and ViT models. | ||