Created: September 14, 2022
Modified: September 14, 2022
Modified: September 14, 2022
superposition
This page is from my personal notes, and has not been specifically reviewed for public consumption. It might be incomplete, wrong, outdated, or stupid. Caveat lector.A -dimensional vector can represent distinct orthogonal features, but due to the weirdness of high-dimensional geometry, it can represent almost-orthogonal features, which we define as feature vectors having a dot product of at most . This allows for networks to represent a 'superposition' of exponentially many potential features in a single vector.
https://transformer-circuits.pub/2022/toy_model/index.html
https://en.wikipedia.org/wiki/Johnson%E2%80%93Lindenstrauss_lemma
An argument I like is that random vectors in high-dimensional space are orthogonal with high probability. (TODO: this argument is currently broken, need to remember/look up what's missing).
- If we treat each dimension as standard normal with variance , then in high dimensions we'll get unit vectors with high probability.
- The dot product of two independent normal vectors is a sum of terms, each of which has mean zero and variance , so the sum itself has mean zero and variance . The probability of the sum being far from zero can be bounded by various concentration inequalities; naively we appeal to the central limit theorem to show that the sum is approximately normally distributed and thus
- We want to make this event so unlikely that, given an exponential number of vectors, and considering all pairs, a union bound still gives a low probability that any of the dot products has magnitude . Summing over all events gives but this is not a useful bound since we are likely to want to take much smaller than 2. So something in this argument needs to be strengthened.