| ▲ | scottlamb a day ago | ||||||||||||||||||||||||||||
There are lots of reasons to read through source code you never edit or recompile: security audits, interoperability, learning from their techniques, etc. And I think many of those same ideas apply to seeing the training data of a LLM. It will help you understand quickly (without as much experimentation) what it's likely to be good at, where its biases may be, where some kind of supplement (transfer learning? RAG? whatever) might be needed. And the why. | |||||||||||||||||||||||||||||
| ▲ | vova_hn2 16 hours ago | parent | next [-] | ||||||||||||||||||||||||||||
> security audits If you are unable to run the multimillion training, then any kind of security audit of the training code is absolutely meaningless, because you have no way to verify that the weights were actually produced by this code. Also, the analogy with source code/binary code fails really fast, considering that model training process is non-deterministic, so even if are able to run the training, then you get different weights than those that were released by the model developers, then... then what? | |||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||
| ▲ | oscarmoxon a day ago | parent | prev | next [-] | ||||||||||||||||||||||||||||
Agree, this feels like a distinction that needs formalising... Passive transparency: training data, technical report that tells you what the model learned and why it behaves the way it does. Useful for auditing, AI safety, interoperability. Active transparency: being able to actually reproduce and augment the model. For that you need the training stack, curriculum, loss weighting decisions, hyperparameter search logs, synthetic data pipeline, RLHF/RLAIF methodology, reward model architecture, what behaviours were targeted and how success was measured, unpublished evals, known failure modes. The list goes on! | |||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||
| ▲ | kazinator 19 hours ago | parent | prev [-] | ||||||||||||||||||||||||||||
Security audits, etc, are possible because binary code closely implements what the source code says. In this case, you have no idea what the weights are going to "do", from looking at the source materials --- the training data and algorithm --- without running the training on the data. | |||||||||||||||||||||||||||||