▲ | bc569a80a344f9c 3 days ago | ||||||||||||||||||||||||||||||||||||||||||||||
An interesting follow-up is using various xAI (explainable AI) techniques to then investigate what features in an image the classifier uses to make its decisions. Saliency maps work great for images. When I was playing around with it, the binary classifier I trained from scratch to distinguish cats from dogs ended up basically only looking at eyes. Enough images in the dataset featured cats with visible, open eyes, and the vertical slit is an excellent predictor. It was an interesting lesson that also emphasized how much the training data matters. | |||||||||||||||||||||||||||||||||||||||||||||||
▲ | cco 3 days ago | parent | next [-] | ||||||||||||||||||||||||||||||||||||||||||||||
ExAI feels like a better shortening, both for clarity and given that xAI is a company already. | |||||||||||||||||||||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||||||||||||||||||||
▲ | vismit2000 3 days ago | parent | prev | next [-] | ||||||||||||||||||||||||||||||||||||||||||||||
▲ | krackers 3 days ago | parent | prev [-] | ||||||||||||||||||||||||||||||||||||||||||||||
This article seemed really basic, no insight other than "it learns the high dimensional manifold on which cat images lie, thus separating cats from non-cats" (not that simple explanations are bad, but Quanta articles seem to be getting more watered down over time). The real question is whether we can get some insight as to how exactly it's able to do this. For convolution neural networks it turns out that you can isolate and study the behavior of individual circuits and try to understand what "traditional image processing" function they perform, and that gives some decent intuition: https://distill.pub/2020/circuits/ - CNNs become less mysterious when you break them down as being decomposed into "edge detectors, curve detectors, shape classifiers, etc." For LLMs it's a bit harder, but anthropic did some research in this vein. |