| ▲ | zahlman 3 hours ago | |||||||
That sounds like the sort of thing I'd expect them to be good at. What goes wrong? | ||||||||
| ▲ | CamperBob2 3 hours ago | parent [-] | |||||||
I edited the post with a speculation, but it's just a guess, really. In the training data, different units are going to share near-identical grammatical roles and positions in sentences. Unless some care is taken to force the embedding vectors for units like "pounds" and "kilograms" to point in different directions, their tokens may end up being sampled more or less interchangeably. Gas-law calculations were where I first encountered this bit of scariness. It was quite a while ago, and I imagine the behavior has been RLHF'ed or otherwise tweaked to be less of a problem by now. Still, worth watching out for. | ||||||||
| ||||||||