| ▲ | nee1r a day ago | |||||||
real | ||||||||
| ▲ | sheepscreek 5 hours ago | parent [-] | |||||||
Are you guys affiliated with Meta’s ex-CTO in any way? I remember he famously implied that LLMs hyped. The demos are very impressive. Does this use an attention based mechanism too? Just trying to understand (as a layman) how these models handle context and if long contexts lead to weaker results. Could be catastrophic in the real world! | ||||||||
| ||||||||