Remix.run Logo
tripplyons a day ago

Is that where you approximate a partial derivative as a difference in loss over a small difference in a single parameter's value?

Seems like a great way to verify results, but it has the same downsides as forward mode automatic differentiation since it works in a pretty similar fashion.

yobbo a day ago | parent [-]

Yes, the purpose is to verify the gradient computations which are typically incorrect on the first try for things like self-attention and softmax. It is very slow.

It is not necessary for auto-differentiation, but this project does not use that.