▲ | int_19h 4 days ago | |
> We actually perfectly understand both how the weights arise and how they result in the model's outputs If we knew that, we wouldn't need LLMs; we could just hardcode the same logic that is encoded in those neural nets directly and far more efficiently. But we don't actually know what the weights do beyond very broad strokes. |