| ▲ | HarHarVeryFunny 2 days ago | ||||||||||||||||
Does the way "manifold" is used when describing subsets of the representational space of neural networks (e.g. "data lies on a low-dimensional manifold within the high-dimensional representation space") actually correspond to this formal definition, or is it just co-opting the name to mean something simpler (just an embedded sub-space)? If it is the formal definition being used, then why? Do people actually reason about data manifolds using "atlases" and "charts" of locally euclidean parts of the manifold? | |||||||||||||||||
| ▲ | antognini 2 days ago | parent | next [-] | ||||||||||||||||
It's hard to prove rigorously which is why people usually refer to it as the "manifold hypothesis." But it is reasonable to suppose that (most) data does live on a manifold in the strict sense of the term. If you imagine the pixels associated with a handwritten "6", you can smoothly deform the 6 into a variety of appearances where all the intermediate stages are recognizable as a 6. However the embedding space of a typical neural network that is representing the data is not a manifold. If you use ReLU activations the kinks that the ReLU function creates break the smoothness. (Though if you exclusively used a smooth activation function like the swish function you could maintain a manifold structure.) | |||||||||||||||||
| |||||||||||||||||
| ▲ | griffzhowl 2 days ago | parent | prev | next [-] | ||||||||||||||||
There's a field known as information geometry. I don't know much about it myself as I'm more into physics, but here's a recent example of applying geometrical analysis to neural networks. Looks interesting as they find a phenomenon analogous to phase transitions during training Information Geometry of Evolution of Neural Network Parameters While Training | |||||||||||||||||
| ▲ | youoy 2 days ago | parent | prev [-] | ||||||||||||||||
The closest thing that you may get is a manifold + noise. Maybe some people thing about it in that way. Think for example of the graph of y=sin(x)+noise, you can say that this is a 1 dimensional data manifold. And you can say that locally a data manifold is something that looks like a graph or embedding (with more dimensions) plus noise. But i am skeptical whether this definition can be useful in the real world of algorithms. For example you can define things like topological data analysis, but the applications are limited, mainly due to the curse of dimensionality. | |||||||||||||||||
| |||||||||||||||||