Remix.run Logo
pests 8 days ago

Technically modern LLMs are handicapped on translation tasks compared to the original transformer architecture. The origami transformer got to see future context as well as past tokens.

vouaobrasil 8 days ago | parent [-]

Okay, but I'm not really concerned with the state of the art now with any specific technology, but what will be the state of the art in 20 years.