▲ | petesergeant 5 days ago | |||||||||||||||||||||||||||||||||||||||||||
I wrote a simpler explanation still, that follows a similar flow, but approaches it from more of a "problems to solve" perspective: https://sgnt.ai/p/embeddings-explainer/ | ||||||||||||||||||||||||||||||||||||||||||||
▲ | k__ 5 days ago | parent | next [-] | |||||||||||||||||||||||||||||||||||||||||||
Awesome, thanks! If I understand this correctly, there are three major problems with LLMs right now. 1. LLMs reduce a very high-dimensional vector space into a very low-dimensional vector space. Since we don't know what the dimensions in the low-dimensional vector space mean, we can only check that the outputs are correct most of the time. What research is happening to resolve this? 2. LLMs use written texts to facilitate this reduction. So, they don't learn from reality, but from what humans written down about reality. It seems like Keen Technologies tries to avoid this issue, by using (simple) robots with sensors for training, instead of human text. Which seems a much slower process, but could yield more accurate models in the long run. 3. LLMs holds internal state as a vector that reflects the meaning and context of the "conversation". Which explains, why the quality of responses deteriorates with longer conversations, if one vector is "stamped over" again and again, the meaning of the first "stamps" will get blurred. Are there alternative ways of holding state or is the only way around this to back up that state vector at every point an revert if things go awry? | ||||||||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||||||||
▲ | tamasnet 2 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||||||||
Thanks for sharing this, I'm really enjoying the style and you've clarified some concepts in a clear way. | ||||||||||||||||||||||||||||||||||||||||||||
▲ | visarga 5 days ago | parent | prev [-] | |||||||||||||||||||||||||||||||||||||||||||
Your approach is much more intuitive. I was coming back to say why didn't they show an embedding with categorical/scalar features? |