| ▲ | HerbManic 4 hours ago | |
I have said for a while that we need a sort of big-little-big model situation. The inputs are parsed with a large LLM. This gets passed on to a smaller hyper specific model. That outputs to a large LLM to make it readable. Essentially you can blend two model type. Probabilistic Input > Deterministic function > Probabilistic Output. Have multiple little determainistic models that are choose for specific tasks. Now all of this is VERY easy to say, and VERY difficult to do. But if it could be done, it would basically shrink all the models needed. Don't need a huge input/output model if it is more of an interpreter. | ||