| ▲ | computomatic 6 hours ago | |
I find myself wondering about this though. Because, yes, what you say is true. Transformer architecture isn’t likely to handle negations particularly well. And we saw this plain as day in early versions of ChatGPT, for example. But then all the big players pretty much “fixed” negations and I have no idea how. So is it still accurate to say that understanding the transformer architecture is particularly informative about modern capabilities? | ||
| ▲ | tovej 5 hours ago | parent [-] | |
They did not "fix" the negation problem. It's still there. Along with other drift/misinterpretation issues. | ||