We propose a new architecture to interpret LLM representations, called RQAE (Residual Quantization Autoencoder). RQAE learns features hierarchically, by applying residual vector quantization on the residual stream of an LLM. RQAE has equivalent or better reconstruction than SAEs, and has much higher capacity.
RQAEs reduce the pathological errors of SAEs iteratively. As a result, they naturally solve the issue with feature splitting and absorption in SAEs. Additionally, RQAE can be used in three common interpretability tasks that SAEs are used for: finding model features (i.e. dictionary learning a set of features), activation steering, and concept detection.
If you are entirely new to interpretability research, start from the beginning.
If you care about how the model works, start here.
If you only care about the connection to SAEs, start here.
If you only care about experimental results, start here.
This page is a work in progress. Here are some TODOs that are remaining, in order of priority:
Add graphics to Algorithms for feature distributions.
Add graphs for ablation studies (training runs already completed).
General writing cleanup.
Publish model weights and code to run steering and feature finding.
Host frontend to view monology features.
Complete toxicity detection experiments.
Update the bibliography.
I hope to have all of these todos done by Oct 25.
The post is written very definitively, but is better described as my current thoughts on the project. I'm still thinking through a lot of the conceptual details, although the experimental evidence I have seen so far is very solid. I'm generally optimistic about this approach, and less confident about my explanations for the approach.
Introduction
The purpose of interpretability is to decompose a deep model into human-understandable features. This has led to some incredibly interesting work: my favorites include visualizing what parts of an image a model uses to predict a class and seeing how RL agents identify friends and enemies in a game .
However, Large Language Models (LLMs) may be more complicated to interpret. Text is a much denser medium than visuals when it comes to interpretable features, and LLMs are orders of magnitude larger than any other models weāve ever trained. This work is heavily inspired by transformer-circuits, which has laid out a foundation for how we can begin to interpret LLMs. If youāre new to interpretability, I would recommend reading these works first, but Iāll provide a quick rundown of the core concepts in the sections below.
Other work has explored how different layers of the LLM contribute to different functions (for example, the feedforward layers store facts that the LLM knows ). However, the scope of this work is limited to considering the residual stream of a LLM. In the future, we hope to extend this model to all other layers of the LLM, as has been done with SAEs .
World Models
What is a human interpretable feature, and why would they exist in LLMs? Well, we know that LLMs store world models because they are really good at predicting the next token, and compressing the training data to the extent that they do requires some latent understanding of the world. Ha & Schmidhuber define this more rigorously , but roughly:
There are a large number of things that can and will happen in the world
We observe what happens, and we reason about what happens with a much, much smaller set of ālatentā features.
For example, if we see someone drop a glass bottle, then we expect the bottle to shatter when it hits the ground. This is not because we have observed exactly that person dropping exactly that bottle before. Itās because we have learned a very small set of latent features about physics, which we can use to model the bottle breaking.
This is exactly why LLMs are so exciting - they just donāt have the capacity to store all of the knowledge that we know they have (i.e. to actually memorize the training data), which means that they must also be working with some set of latent features that they can manipulate to solve the next token prediction task!
Linear Representations and Superposition
How do LLMs organize and use these latent features? Thereās a lot of evidence that LLMs describe features as simply directions in space, known as the Linear Representation Hypothesis (LRH). If true, itās a very powerful framework - for example, you can define a causal inner product, and show that vectors which are orthogonal to one another are also causally linked .
Then, a full LLM activation is the sum of multiple atomic interpretable features. For example, we might think of a dog as the āanimalā feature plus the āpetā feature, minus the āfelineā feature. Features also scale differently depending on the subject: a dog will have more of the āsmartā feature than a goldfish.
We formalize the LRH with the definition given here:
Definition 1: A linear representation has the following two properties:
Composition as Addition: The presence of a feature is represented by adding that feature.
Intensity as Scaling: The intensity of a feature is represented by magnitude.
There is one problem with the LRH: the residual stream of an LLM with width $d$ lies in the space $\mathbb{R}^d$. However, there can be at most $d$ orthogonal vectors in this space (any basis you choose), which means that there can be at most $d$ completely unique āfeature directionsā.
In order for the LRH to be true, models must learn features in superposition. This means that features will interfere with each other, and you can not fully separate feature interactions. However, you can also fit exponentially more features in a $d$-dimensional space which are only $\epsilon$-orthogonal to each other!
Thereās one more thing to notice about the LRH: similar features should correspond to similar directions. For example, the ādogā feature should be closer to the ācatā feature than to the āsharkā feature. Thus, when measuring by cosine similarity, we should see clusters forming which correspond to similar features.
Sparse Autoencoders
Sparse Autoencoders (SAEs) try to take advantage of the LRH by learning an overcomplete basis to take features out of superposition . SAEs take inspiration from sparse coding theory (Chapter 7 from Wright and Ma is a great introduction): it suggests that learning a dictionary of features that only fire sparsely (i.e. a very small number of features are required to define each activation) also results in that dictionary being human-interpretable.
A figure taken from Toy Models of Superposition. In this case, the observed model is our LLM, and the disentangled model is what we are trying to learn with a SAE.
Looking back at the World Models section, this actually makes a lot of sense! We can think of the ālargerā model as the set of disentangled latent variables that model the real world. Importantly, the larger model only activates sparsely - this means that any interaction only takes in a few features. Sparse autoencoders try to reconstruct this larger model, in order to move out of superposition.
Definition 2: A sparse autoencoder $S$ of size $n$ is a model that takes in an activation $r \in \mathbb{R}^d$ and does the following:
$$
C(r) = \sigma(W_{in}r + b_{in})
$$
$$
S(r) = W_{out}C(r) +b_{out}
$$
where $\sigma$ is some nonlinearity (usually ReLU), and $W_{out} \in \mathbb{R}^{d \times n}, W_{in} \in \mathbb{R}^{n \times d}$ (for $n \gg d$). $b_{in} \in \mathbb{R}^n, b_{out} \in \mathbb{R}^d$ are bias terms.
S is trained to recover $r$ while inducing sparsity in $C(r)$, meaning that $C(r)$ is zero in most dimensions - common methods to induce sparsity include TopK, L1 loss, or just thresholding.
To explicitly draw the connection to the LRH, a SAE performs the following algorithm:
Begin with some entangled LLM activation $r$, and consider that $W_{in}$ consists of $n$ vectors, which we will call āprobeā vectors.
Compare $r$ against each probe vector by calculating their dot product.
Add a bias to each of these dot products, because some features scale differently than others depending on their average intensities.
Apply a nonlinearity (usually ReLU), to act as a filter for sparsity (by dropping probes that are very unlikely to zero).
There are now $n$ coefficients $C(r) = [c_1, c_2, ā¦]$, which are the intensities of each feature in $r$.
Multiply $C$ by the columns of $W_{out}$, which are the actual feature vectors (directions) that you have learned.
NOTE: Notice that each (probe, feature) pair of vectors are closely linked - the cosine similarity ($\propto$ dot product) that a representation has to a probe directly defines the intensity of the feature. At the beginning of training, the encoder is initialized as the transpose of the decoder so that probes = features.
However, an SAE can only learn a set number $n$ of features, and itās likely that $n \ll N$ for the true number of features $N$ the LLM uses. To give a toy example, letās consider there to be two āground-truthā features $f_1$ and $f_2$ - these features are similar but not exactly the same (e.g. different breeds of dogs). How will the SAE learn these features?
The Feature Hierarchy
The answer to the question posed above is: the model will learn an average direction that will fire weakly for these features - for example, a ādogā feature. It might also learn a few features that weakly activate for related, more general features - for example, a ālivingā feature, or a āpetā feature. Clearly, we want some level of control on the specificity of features that SAEs learn.
There are three widely observed issues with SAEs, that all stem from this intuition:
Feature splitting. If you train a wider SAE, you will notice that more general features split into smaller, more specific features.
Feature absorption. You learn two separate features in your SAE that are describing the same ground-truth feature - as a result, representations with that ground-truth feature are split across the two learned features without any discernable pattern (i.e. the difference between the two features is spurious).
Feature shrinkage. SAEs routinely underestimate the intensity of a given feature. It happens because of the sparsity penalty during training - see this work for a clear example and more. JumpReLUs largely mitigate this issue, but itās related to the first two because it happens due to learning entangled features (an SAE will underestimate a featureās intensity because it wants to account for other features that will interfere).
All three of these issues stem from the same underlying cause: SAE features are not necessarily atomic. They might need to be broken down or grouped together, but itās difficult to tell which is which, and the training objective doesnāt bias the model one way or the other. This begs the question: what is an atomic feature?
A paper from UChicago attempts to answer this question . They split up features in two dimensions: hierarchical features, such as organism -> (plant -> (tree, bush, etc.), animal -> (bird, reptile, fish, etc.)), and categorical features, such as (dog, cat, hamster, etc.). Hierarchical features are organized orthogonally to each other, while categorical features organize themselves into polytopes.
SAEs treats features as if they are all categorical. It can learn hierarchies of features (for example, it can learn separate features for organism, plant, animal, etc.) - but there is nothing in the architecture that encourages it to learn such features. My best guess is that it learns hierarchy based on the frequency of the concepts in the training data alone, since this work finds that SAEs learn more granular features when trained on different datasets. However, this is still an open question.
Brainstorming a New Architecture
Consider a new architecture that did learn features hierarchically. What would that look like? It should assign features to a hierarchy, with parent and child features, such that any time a feature is activated, itās parent will also be activated, and zero or one of itās children will be activated. Then, we could address the three issues presented in the previous section:
If a feature splits, then we define the base feature as higher in the hierarchy, and the split features as lower in the hierarchy.
Feature absorption is a result of going deeper in the hierarchy than you want to. If you notice two features that should be absorbed, you should ignore them and only consider their parent feature.
Feature shrinkage should not happen, because features at a given layer in the hierarchy will only ācompeteā with other features on the same level. That is, the intensity of a parent feature shouldnāt depend on the intensity of itās children, and vice versa. Since exactly one feature will be activated at each layer of the hierarchy, there is no incentive to account for interference.
RQAE
Finally, we introduce the proposed architecture, RQAE (Residual Quantization Autoencoder). This architecture follows the model described in the previous section, and learns features hierarchically. Check out Figure 1 for an illustrated version.
The core idea behind RQAE is to use residual vector quantization (RVQ) to autoencode the LLM representation. If you arenāt familiar with RVQ, here is a great explanation. The primary difference is that we inject a linear in/out layer between each quantization step, which allows the model to iteratively choose different subspaces of the representation space to quantize.
We use a quantization variant of FSQ , that uses hyperspheres instead of hypercubes to define codebooks. There are two benefits of using FSQ:
There are no issues with codebook collapse. Most codebooks are utilized fully, which is especially important when we have many layers.
It restricts the geometry of subspaces in the representation space to ellipsoid objects only (hypersphere + affine transform). Since ellipsoids can approximate convexCategorical feature polytopes don't necessarily have to be convex, so this claim is somewhat dubious - but in practice, it seems to be okay. polytopes well , subspaces are encouraged to take the shape of the categorical features mentioned above.
Figure 1. An overview of RQAE. This diagram does not include biases, which are present at each linear layer.
RQAE fits the definition from above well. For any given LLM activation, every layer returns exactly one codebook. The subspaces that later layers span are dependent on earlier layers (not strictly, as we will see later), encouraging a hierarchy between layers. We define the learned features to be the entries of the codebooks after they are projected by the linear out layer (see Figure 1). Thus, a RQAE model with $n$ layers and $c$ codebooks per layer learns $c \times n$ unique features, but can represent $c^n$ different activations.
A Close Resemblance to SAEs
Itās hard to propose a new architecture, when the existing architecture already has a large body of work and plenty of empirical evidence (for example, Gemmascope used $O($training compute of GPT-3$)$ to train SAEs). Thus, we draw a connection between the layers of a RQAE and a SAE.
Conceptually, RQAE has the same parts as the SAE algorithm mentioned above: linear in layers are āprobesā, groups of intensity coefficients are constrained to lie on the hypersphere, and features are the columns of the linear out layers. However, we can go a step further and show mathematically that these two are the same:
Lemma 1: A single layer of RQAE can be equivalently defined as an SAE.
Proof: Still a WIP. Will follow this work closely
In practice, notice that in Figure 1 we use a much smaller codebook dimension. This is needed for quantization (especially FSQ, whose codebook size grows exponentially with respect to codebook dimension). However, this is not a concern, because there is a large body of work that suggests LLM features and feature manifolds are represented in very low dimensional subspaces (i.e. rank deficiency).
For clarity, here is a table of comparison between SAEs and RQAE:
SAE
RQAE
Method
Learn a set of "probe" vectors, which match representations by cosine similarity. Each probe vector has a corresponding "feature" vector, whose intensity is the dot product of the probe and representation.
Learn a low-dimensional ellipsoid that best covers the representation space. Quantize the surface evenly, where each quantized point is a feature. Project the representation onto the ellipsoid and choose the closest quantized point. Now, consider the difference between the representation and its projection. Repeat. (See Fig 1)
Feature Splitting
Train SAEs of different widths.
Consider features at different levels of the hierarchy.
Interpreting Features
Run the model on a test dataset. Look at the max activating tokens for each feature.
Choose a token (or a small set of tokens) as a query. Look for tokens in the test dataset that are the most similar to the query.
Steering
Encode the representation with the set of probes. Then, inflate the intensity of the feature that you want to steer. Decode and continue running the LLM.
Learn a distribution for the feature at each level of the hierarchy. Then, force the representation to match the learned distribution for a subset of layers.
Comparison between SAEs and RQAE. Note that this table covers information from later sections as well.
Modeling Pathological Errors
This work describe how SAEs create pathological errors - that is, the residual between SAEs and original activations are uniquely important for reconstruction (measured by cross entropy loss difference), compared to completely random residuals the same distance away. Exactly why this happens is still an open question - but the fact that it happens is important!
Itās reasonable to assume that features uniquely impact reconstruction - for example, missing critical features in the reconstruction of an activation will result in the model having a harder time predicting the next token, compared to adding random noise that simply adds more interference between all features but doesnāt remove any features. Then, the fact that SAEs create pathological errors could mean that they are missing features! This also fits the empirical observation of feature splitting. When a feature splits, you are adding learned āsub-featuresā to an existing feature (for example, you are adding the different breed characteristics to a ādogā feature). These sub-features are the missing features that impacted reconstruction.
This provides a good motivation for RQAE. Specifically, a single layer of RQAE is equivalent to an SAE, as shown above. The next layer of RQAE then learns on the residuals of the first layer, which we assume to have āmissing featuresā, since reconstruction is not perfect. Applying this iteratively suggests that RQAE features correspond to the same āsub-featuresā that split existing SAE features.
Design Decisions
RQAEs can be narrowly defined with the framework of SAEs given the right hyperparameters, but the actual implementation of RQAE can be very different. In this section, weāll describe notable design decision made while implementing RQAE. Hereās the default model we will reference throughout the rest of the work:
Parameter
Value
Description
LLM
Gemma 2 2B/9B
Base model to interpret
Layer
Residuals after the center layer (12 for 2B, 21 for 9B)
What layer of the residual stream to train on
Training Data
FineWeb-Edu
Dataset used to run the LLM on
Test Data
pile-uncopyrighted (Neuronpedia subset)
Dataset used for analysis
num_tokens
1B
How many tokens to train on
context_length
128
Context length to train on
num_quantizers
1024
Number of residual quantizer layers
codebook_dim
4
Dimension of codebook
codebook_size_per_dim
5
Size of codebook, per codebook dimension$^\star$
Table 1. Default parameter values for RQAE.
Hypersphere FSQ: We use a variant of FSQ, that uses hyperspheres instead of hypercubes to define codebooks. This way, the model only has one code that fully activates each feature, making it easier to interpret features individually.
Unconstrained Decoder Norm: Quantization doesnāt have an equivalent of intensity, compared to SAEs. Thus, in contrast to SAEs, we do not restrict the decoder features to have unit weight norm.
Normalized Activations: Before passing the LLM activations into RQAE, we normalize them. For simplicity, we use the final RMS norm layer of the LLM. Note that this means the activations donāt actually have unit norms, but rather they have the same mean and variance This is different than SAEs, which normalize activations by averages over the dataset. This wasnāt done intentionally, but didnāt seem to affect training - itās future work to investigate this further.
Training and Test Dataset: We use the FineWeb-Edu dataset, because we expect the focus on educational content to allow RQAE to learn more interesting interpretable features. We use the same subset of monology/pile-uncopyrighted that Neuronpedia uses to interpret Gemmascope features.
128 Context Length: We found many prior works to use this context length, but some works (notably Gemmascope ) to use longer context lengths. We stick to 128, because it allows for faster experimentation, but we may introduce future work extending this limit. Notably, Gemmascope found that reconstruction loss is relatively constant after the first several tokens.
1B training tokens: Most prior work trains for longer (by 4-16x). Indeed, we do see that MSE loss is still decreasing at 1B tokens. However, the loss hits an inflection point even as early as 200M tokens (see Figure 3a), and similar to above, we wanted to prioritize faster experimentation.
For all other parameters, we provide empirical evidence in the Ablation section, describing how they impact performance.
Properties
First, we look at the properties of RQAE features, to validate assumptions we have based on the model architecture.
Figure 2a.(left) Distribution of pairwise cosine similarities between all learned features. (right) L2 norm of features across RQAE layers.
Figure 2a shows the distribution of pairwise cosine similarities between all learned features. Almost all features are $\epsilon$-orthogonal with $\epsilon=0.2$. It also shows the L2 norm of features across RQAE layers. As expected, earlier layers have larger features than later layers, since we donāt constrain the decoder weight norm. From this, we can tell that earlier layers choose more āconfidentā directions - i.e. the majority of MSE loss is concentrated in the first few layers.
Figure 2b.(left) Distribution of codebook usage across 1M test tokens for the first 16 layers. (right) The same graph, but displaying the coefficients for each feature that the codebook imposes (e.g. codebook $302 = [0, 0, -1, 0]$).
FSQ promises better quantization, because it utilizes codebooks fully. In RQAE, this would mean that we are finding the most informative ellipsoids to cover the representation space. Figure 2b shows the codebook usage for a set of tokens. We can see that, on average, FSQ is utilizing all of its codebooks across layers, for a large number of tokens. The right graph also shows the coefficients that features are multiplied by. We can see that most feature weights are $0$ - suggesting that the four ellipsoid axes are generally a good sparse representation of the data.
Figure 2c. Both images display the bottom $16$ feature coefficients, sorted by variance, for a set of tokens. (left) Tokens are chosen from the max activating tokens for a given Gemmascope SAE feature. (right) Tokens are chosing randomly from the test set.
Finally, letās consider how RQAE features can be used to interpret LLM representations. In Figure 2c, we compare features learned on random tokens, and learned on the max activating tokens for a random feature that the equivalent Gemmascope SAE found. We perform the following algorithms:
Encode the representation with RQAE, and look at all of the feature coefficients (i.e. codebooks on the hypersphere) individually.
Sort feature coefficients by variance across all tokens.
Look at the bottom $16$ coefficients. This corresponds to features that all tokens share similarly.
We can see that the Gemmascope feature tokens are much lower variance than random tokens. This suggests that the features learned by RQAE can be used to partition tokens that share similar features.
Reconstruction
In most previous work, reconstruction is seen as a proxy for capacity. That is, a model with lower reconstruction loss can hold more features - and itās implied that these features will also be interpretable. As a result, we examine how good RQAE is at reconstruction.
Figure 3a. MSE Loss and Cross Entropy Loss Difference for RQAE trained on Gemma 9B and 2B models. Both RQAE models are trained with num_quantizers=1024.
Figure 3a shows the MSE loss and cross entropy loss difference for RQAE trained on Gemma 9B and 2B models during training. Clearly, RQAE has enough capacity to reconstruct the original LLM activations faithfully, and it seems that further training will continue to decrease reconstruction loss (future work).
We also notice that cross entropy difference saturates relatively quickly, while MSE decreases more slowly (check out Ablations for a more detailed experiment). This is likely related to a recent finding that the reconstruction error in SAEs are pathological - the residual error affects reconstruction more than an equally large random residual.
Figure 3b. Reconstruction loss of RQAE vs SAE with varying width and L0 for Gemma 2 2B. Note that the same RQAE model is used here (num_quantizers=1024), and reconstruction is stopped early, rather than different RQAE models trained with different num_quantizers.
We can also compare directly against the equivalent SAEs in Gemmascope. Notice that RQAE has much lower reconstruction loss when using all of its quantizers, even compared to SAEs with high L0. Of course, these models were trained on different data. However, the test data is chosen independently from either model, and the performance difference is significant, so we believe that the conclusion still holds.
It might be tempting to claim that you should only compare reconstruction loss where num_quantizers = $L_0$, since RQAE chooses a feature per quantizer for each representation. However, this ignores two important distinctions:
There are significant geometric constraints placed on RQAE features in any given quantization layer: not only must they exist on an low-dimensional ellipsoid, but they also choose from a small set of quantized values on that ellipsoid.
At each layer, a feature must be chosen, as opposed to SAEs, where features can be used in arbitrary combinations.
We donāt think there is a clear equivalence between num_quantizers and $L_0$ (maybe there is some connection between num_quantizers and $L_0$ of a meta SAE, but this is future work), and suggest that you only consider the actual representational capacity between the two models.
Interpretability
The previous section showed that RQAE has much higher capacity than SAEs - however, this doesnāt mean that the learned features are any more interpretable, which is really what we care about. To interpret features, weāll follow the most naive approach first: look at tokens in a dataset that quantize to the same codebooks.
Figure 4a. Three examples of clusters of tokens with the same codebooks, starting at layer 0 and splitting on layer 1. The green tuple in each box represents the codebooks used. For example, $(198)$ means that this box includes all tokens from the test set that use codebook $198$ in layer $0$. $(198, 32)$ means that this box includes all tokens that use codebook $198$ in layer $0$, and codebook $32$ in layer $1$.
Figure 4a shows three examples of clusters of tokens with the same codebooks, starting at layer 0 and splitting on layer 1. There is some noise, but generally, we see that tokens using the same codebooks share similar interpretations. This is exciting: it means that the points chosen on the ellipsoid are actually representing interpretable feature directions!
Figure 4b. An example of how different codebooks at layer 1 can interfere with each other. In this example, all four boxes include tokens relating to politics and healthcare.
Unfortunately, Figure 4a was mostly cherry-picked to show clear feature examples. Figure 4b shows an example of how different codebooks at layer 1 can interfere with each other. In this example, all four boxes include tokens relating to politics and healthcare, with no discernable pattern separating them. This is the same problem of feature absorption that SAEs have!
Figure 4c. Instead of viewing specific codebooks, we can view all codebooks within a cosine similarity threshold to an existing codebook.
We can mitigate feature absorption by remembering that all features in a given layer must lie on an ellipsoid. This means they arenāt entirely independent - so we should be considering clusters of codebooks at a time.
Figure 4c shows all codebooks within a cosine similarity threshold to a given codebook. For example, there are $16$ codebooks within $0.9$ cosine similarity distance to codeboook $568$, so we can consider all tokens using any of these codebooks. By doing this, we see some more general patterns. High cosine similarities result in more similar tokens, negative cosine similarities result in dissimilar tokens, and so on.
Issues with a Strict Hierarchy
We would like to note that the method described in the above section is very naive. There are a lot of issues with the approach presented:
There are exponentially many combinations of codebooks as the number of layers increases. For example, the two layers that we visualized in this section (0 and 1) have $625^2 = 390625$ unique combinations. Since we saw in Properties that codebooks are utilized fully, this means that we will also need exponentially more test data to continue to interpret clusters of tokens with enough statistical significance.
Layers 0 and 1 have the highest magnitude features (Figure 2a). However, they still are constructed from only two low dimensional ellipsoids. This means that at best, we are considering $8$ dimensions of the original $2034$-dimensional representation space.
Thereās still a lot of noise in the examples given - Figure 4a has clean deliniations between features, but Figure 4c has overlapping tokens even after using cosine similarity to cluster codebooks. Itās hard to make strong interpretability claims with so much noise.
The above issues stem from a fundamental constraint with how we have been viewing RQAEs: as a strict hierarchy of features. In reality, this doesnāt make sense: many (most) features in a LLM donāt lie in a single $8$-dimensional space, so we expect that the first two layers of RQAE should provide relatively little information about them. Figure 2c hints at how we can relax this constraint, because notice that the lowest variance features are spread throughout layers. Figure 3b also suggests the same thing: later quantization layers are still essential to reconstruction, and it does not seem to follow an exponential decay that you would expect with a strict hierarchy. This leads us to a new interpretation of RQAE:
The layers of RQAE do not strictly form a hierarchy of features in order. Each human-interpretable feature (or feature manifold) chooses a different subspace to exist in, and RQAE layers choose the best set of subspaces to explain them. As a result, each human-interpretable feature will have a different hierarchical ordering of subspaces that explain it.
We can find this hierarchy for a given feature by following Figure 2c - finding RQAE features with the lowest variance across examples of the human-interpretable feature. In the next section, we will examine how this insight can be used in practice.
Algorithms
After an SAE is trained, itās features are interpreted in the following way:
Choose a large test dataset, with a variety of features you care about.
Run the SAE on the dataset, and select the max-activating tokens for each SAE feature.
Interpret the set of max activating tokens for each SAE feature, either manually or with another LLM (e.g. AutoInterp).
Use these interpretations for downstream tasks, such as steering.
There is an implicit assumption with this method of interpreting features: that there are enough tokens in the test set that will activate each feature. If this is not the case - if some features are not activated by enough of the tokens in the test set - then this method might lead to spurious interpretations. As a result, the statistical significance of the interpretations on a test dataset should be carefully considered.
We propose a different approach to interpreting features with RQAE in this section. Because we know that the capacity and reconstruction error of RQAE is much higher than that of SAEs, we assume that all tokens in a test set will be faithfully represented by the RQAE model, and thus all test tokens provide some information about what features they belong to.
Setup
The test set we use is a subset of monology/pile-uncopyrighted - the same subset that is used to interpret Gemmascope SAEs in Neuronpedia. As a result, we can make direct comparisons between the interpretability of Gemmascope SAEs and our RQAEs. This dataset consists of $36864$ sequences of $128$ tokens each, and a variety of human-interpretable features.
We use the same default RQAE model described above. We evaluate on Gemma 2 2B.
Feature Distributions
Redefining the hierarchy showed us that different human-interpretable features can rely on different layers of RQAE. As a result, human-interpretable features are consistent only on some layers (Figure 2c), and for those layers, the variance of feature coefficients is low. We extend this insight by learning feature distributions.
Assume that you have some set of tokens, which share some human-interpretable feature. Then, for each RQAE layer, we can learn a distribution of the feature coefficients (i.e. the codebooks) for that layer. The human-interpretable feature is represented as the set of distributions for each layer. Then, we hypothesize that other tokens which fit most of these distributions also share the same human-interpretable feature.
We use the von Mises-Fisher distribution (vMF) to model the distribution of feature coefficients for each layer, since each codebook is a point on a hypersphere. This distribution is the equivalent of a Gaussian on the surface of a sphere.
// TODO: Add a graphic
Finding New Features
In contrast to SAEs, which look for the max activating tokens in a dataset to define a feature, we propose feature finding by running a query and ranking results. Specifically, we begin by choosing a token from the test set, and:
Encode the token with RQAE.
Calculate the cosine similarities for each layer between all other tokens in the test set and the query token.
Weight the similarities by the L2 norm of each RQAE layer.
Sum the weighted similarities for each token to get a total similarity score.
Return the tokens sorted by their similarity score.
Figure 4a. Using queries to find features. The query token is in bold. For each example, a subset of the top 100 ranked tokens are chosen. This subset chooses the first instance of each token in the ranked list.
Feature Splitting
We can extend the above algorithm to work with multiple query tokens by learning a vMF distribution for each layer.
Encode a set of query tokens with RQAE.
Learn a vMF distribution for each layer.
For each token in the test set, calculate the pdf of the tokenās codebooks under each layerās vMF distribution.
Weight the pdfs by the L2 norm of each RQAE layer.
Return the tokens sorted by their similarity score.
This unlocks much higher specificity when defining features. For example, we can choose subsets of the top 100 ranked tokens to split features into sub-features.
Figure 4b. Splitting a feature from Figure 4a using subsets of it's examples.
Activation Steering
We can also steer activations using vMF distributions. Assume that you have found some feature using the algorithm described above. For a given hidden state:
Encode the hidden state with RQAE.
Calculate the pdf of the hidden stateās codebooks under each layerās vMF distribution.
Steer (i.e. regress towards the mean) each codebook towards the mean of the vMF distribution in the corresponding layer. The strength of steering is determined by the āstandard deviationā of the vMF distribution (i.e. $1/\kappa$) - the lower the standard deviation, the more the hidden stateās codebook is steered towards the mean.
Repeat the process for all hidden states during generation. To control steering, you can introduce two parameters:
top_k: Only steer the top k layers, as measured by $\kappa$. This means that we only steer layers with the lowest variance.
strength: Scale $\kappa$ by some amount. We use exponential scaling between 0 and 1 (code to be released soon). When strength is 0, all distributions become uniform. When strength is 1, all distributions turn into point masses on the mean. A strength of 0.5 means that the distribution remains approximately the same.
Letās look at some examples.
Steering on Gemma 2 2B using the "Bodies of Water" feature from Figure 4a. Use the tabs above to view outputs with/without steering.
An interesting animal is the African elephant. The elephant is the largest land animal on the planet. The elephant is a large animal, but it can also be quite small. The smallest elephant is the dwarf elephant, which is only about 2 feet tall. The African elephant is the largest elephant on the planet. It can weigh up to 14,000 pounds and stand up to 13 feet tall. The elephant is an amazing animal. It is a very intelligent animal and can learn things very quickly. The elephant is also a very social animal. It lives in large herds and is very protective of its young. The elephant is also a
An interesting animal is the European rabbit, which is a symbol of the European continent. It is a member of the Leporidae family, which includes the hare and the hare. This animal is considered a wild animal, but it is a domestic animal in many countries.
The European rabbit is a small animal that is not dangerous to humans. It is one of the most common animals in the world. It has many uses, including food, clothing, and shelter.
The European rabbit is a small, furry animal that is native to Europe. It is also known as the European hare or the European rabbit. It is a member of the Leporidae
An interesting animal is the Psammodromus sceleratus, a desert lizard that lives in the sand dunes of the Sahara Desert.
The Psammodromus sceleratus is a burrowing lizard that has a special adaptation to live in the sand dunes of the Sahara Desert.
It is a nocturnal lizard that is very well camouflaged in its natural habitat, which is the sand dunes of the Sahara Desert.
It is a small lizard that can reach a length of 10 cm, and it has a very interesting color pattern that allows it to blend in with the sand.
An interesting animal is the giant earthworm.
This is a worm that can grow up to 3 meters long. It can be found in Australia, New Zealand, South America and Africa.
It is a very strange looking creature. It has an elongated body, with a segmented structure. It has ten pairs of legs, which are used for locomotion.
It also has a mouth, which is used for feeding. It is a scavenger, and it feeds on decaying matter.
The giant earthworm is a very interesting animal. It is a great example of how evolution can create new and strange creatures.
An interesting animal is found in the desert. It is a small, dark-colored snake with a long tail. The snake is usually found in the desert, but it can also be found in other habitats. The snake is a carnivore and feeds on small animals. The snake is also a venomous snake and can be dangerous.
The desert snake is a species of snake that is found in the deserts of North America, Africa, and Asia. The desert snake is a small, slender snake that is about 12 inches long. The desert snake is a brown or gray color with a white or yellow belly. The desert snake is a venomous snake and
An interesting animal is the Olinguito, a small cat that lives in the mountains of Colombia, Ecuador, and Peru. It is the only member of the family Bassaridae, the only cat family in the world that is not in the Felidae family.
The olinguito is a small cat that is about the size of a house cat, but it is a bit more compact. It is a little less than a foot tall and weighs about 1.5 pounds. Its fur is a light brown color with dark spots on it.
The olinguito is
An interesting animal is the Japanese Giant Hornet. It is a species of wasp and is also known as the Japanese giant hornet or murder hornet.
The Japanese giant hornet is a social insect and is one of the largest insects in the world. It is a large wasp that can grow up to 2.5 inches long and has a wingspan of 3 inches.
The Japanese giant hornet is a social insect, meaning that it lives in large colonies with several queens and workers. The colonies can consist of up to
An interesting animal is the giant Australian octopus (Octopus tetricus). It is a member of the octopus family, the cephalopods, and has the largest brain of any invertebrate. Its brain is about the size of a human brain. The giant Australian octopus is a marine animal that lives in the ocean. It is found in the waters of Australia, New Zealand, and the Pacific Ocean. The giant Australian octopus is a large animal, measuring up to 2 meters (6.6 feet) in length. It has a body that is shaped like a football, with eight arms. The arms are long and thin, and they are used to catch prey.
An interesting animal is the sea cucumber. Its appearance is very unusual, but it is quite common. It is found in the oceans of the world, in the seas, and in the waters of the continental shelf.
The body of a sea cucumber is soft, but not very elastic. It is covered with small, not very hard plates. This shell is composed of calcium carbonate.
The body of a sea cucumber can be very long, and a very large one can be up to 1.5 meters long. The length of a sea cucumber can also vary. It depends on the species, as well as on the sex of the animal. In
An interesting animal is a giant squid. It is the largest invertebrate. The body length of an adult is approximately 12 meters, and the weight is 200 kg. The body of a giant squid is very similar to the body of a human. The size of its body is determined by the presence of an organ called a mantle. This organ is a bag that is located between the body and the beak. The mantle is a very important organ for the squid. It is responsible for the functioning of the organism.
The giant squid is a large animal that lives in the ocean. It is a carnivore and eats fish,
An interesting animal is the sea turtle. It is a reptile, but it lives in the ocean. It is a very slow animal. It canāt swim very fast. That is why it is called a turtle. But it is a very good swimmer. It can swim very fast. It has a shell, which is like a roof. It is made of bone, but it is very hard. The shell is like a helmet, and it is very strong. It is also very protective. It has a long tail. It can also swim very fast. It is very fast and it has a long tail.
It is a very interesting animal.
An interesting animal is the Platyhelminthes, which is also known as the flatworms. These flatworms have a single, simple cavity, and they are also known as acoelomates. The animals that belong to this phylum are bilaterally symmetrical, and they are also found in the phylum.
The animals are also known as platyhelminth, and they are also known as the flatworms. The animals are also known as the turbellarians, and they are also known as the turbellarians. The animals are also known as the
An interesting animal is the tardigrade, or "water bear," which is a microscopic arachnid. It can survive in the vacuum of space. It can withstand temperatures as hot as 140 degrees Celsius and as cold as minus 200 degrees Celsius. It can survive without food for 10 years. It can survive in the vacuum of space. It can withstand temperatures as hot as 140 degrees Celsius and as cold as minus 200 degrees Celsius. It can survive without food for 10 years.
The tardigrade is a microscopic arachnid. It is a microscopic arachnid. It has been found in
An interesting animal is the Starfish. They are often found attached to rocks, and they are a very important part of the ecosystem. They are also very unique in that they can regenerate their body parts. As a result, if a starfish loses a limb, it can grow a new one. The starfish is also known as the sea star. They are found in the ocean.
What is a starfish?
Starfish are invertebrates that live in the ocean. They are not fish, but they do have a similar appearance. Starfish are typically found in the northern hemisphere, but they can also
An interesting animal is a sea cucumber. It is a marine animal that is characterized by a soft body. In the water, it is an almost colorless creature, but when it is threatened, it becomes completely transparent.
The animal is a hermaphrodite, but the females are larger than the males. The sea cucumber is a filter feeder, which means that it filters out food from the water. It is a filter feeder that is able to do this because of the tube that is located on its mouth.
The tube is called a proboscis and it is used to filter out food. The sea cucumber is a filter feeder that is able to do
An interesting animal is the Spiny-tailed iguana, or Iguanaserpentina. It is a species belonging to the Squamata family, more specifically to the Iguanaidae subfamily. It is a reptile, but it has some characteristics of amphibians, as it can also breathe through its skin. It is a cold-blooded animal, and its body temperature is always the same
Notice that in this case, a very low top_k is used. We have found in practice that this is dependent on the feature being steered (similar to SAEs). Generally, top_k is a coarser parameter, that you can set as the maximum value before outputs start degenerating. Then, use strength to fine-tune the strength of steering.
Steering on Gemma 2 2B using a "Compound Word with Hyphen" feature, whose queries are defined below. Use the tabs above to view outputs with/without steering.
Queries
An interesting thing happened in the last few days. I was reading a book and I came across a term that I had never heard before. I looked it up and found that it was a term that was used in the 19th century, and I found it interesting and intriguing. I was wondering if anyone else had heard of this term and what it meant.
The term is called a ādisjunctionā and it means that something is not true. It is a term that is used in the context of a philosophy of science, and it is used to describe a situation where two things are not true.
The term ādisjunctionā
An interesting thing happened the other day. I was in the middle of a meeting and someone asked me the question āWhat does your job entail?ā
Iāve been asked this question before, but Iāve never been able to answer it. Itās a very difficult question to answer because it doesnāt give any insight into what I do.
Iāve tried to answer it before, but Iāve always felt like I was leaving something out.
Iāve tried to answer it in a way that would make people think Iām doing something really important. Iāve tried to answer it in a way that would
An interesting thing happened in the first few days of the new year.
I was in the middle of a very busy week, and I was so busy that I wasnāt able to sit down and watch the first few episodes of the new season of The Mandalorian.
Iāve been a fan of the show since it first aired, and Iāve been following it closely ever since.
But I was so busy that I forgot to watch the first episode of the season.
Iād always been interested in the show, but I wasnāt sure if I was going to be able to watch the first episode.
An interesting thing happened in the last couple of weeks of the 2013 season. The Cubs started to play better baseball.
They were winning games, they were winning series, and they were winning the division.
They had a lot of things to be happy about, and they should be.
They had a lot of things to be happy about, and they should be.
But they also had a lot of things to be unhappy about, and they should be.
This is a team that has not won a playoff game since 1945.
This is a team that has not won a World Series since
An interesting thing about the Star Wars franchise is that it has managed to keep its fanbase engaged for over 40 years now. Itās not just the original trilogy or the prequels that have kept the audience hooked, but also the spin-offs like Rogue One and The Mandalorian.
One of the most popular characters in the entire franchise is the bounty hunter Boba Fett, played by Temuera Morrison. He first appeared in The Empire Strikes Back and then made a comeback in The Book of Boba Fett.
In a recent interview, Temuera Morrison talked about his experience working
An interesting thing happened to me yesterday: I got a call from a man who was very distressed.
He had a 16-year-old daughter who was going through a difficult time with her boyfriend. The boyfriend was a bully. He had been harassing the girl for some time, but it was getting worse. He was threatening to kill her.
The daughter was afraid to tell her parents. She was afraid of the boyfriend, and she was afraid of her parents.
I had a long talk with the girlās father, and we decided to get the police involved.
The young man was arrested and charged with assault and battery
An interesting thing happened while I was researching this post. I found a lot of interesting things about the history of the term āsissyā. I had no idea that it originated from the word āsissieā. I also did not know that the word āsissyā has been around for a long time. It has been around since the 1500ās, which is almost 500 years ago.
The word āsissyā is used to describe someone who is feminine, weak, or cowardly. The word is used to describe a boy or man who is effeminate. The word has
An interesting thing about the 2020 election is that it is the first election in which the race for president is being decided by a majority of the voters.
In 1960, John F. Kennedy defeated Richard Nixon by a margin of just 110,000 votes (0.2 percent) in a three-way race. In 1972, Richard Nixon defeated George McGovern by a margin of 520,000 votes (0.7 percent) in a two-way race. In 1980, Ronald Reagan defeated Jimmy Carter by a margin of
An interesting thing to observe about the two-volume set under review here is that, like the proverbial dove of the title, it is a rather small volume. It is a collection of essays that is as well-written and well-edited as the best-of-the-best in the genre, but it is not an academic-type-of-deal. It is not a scholarly-style-deal. It is, in a sense, an academic-style-deal, but a rather unusual one. It is a little like the 19th-century-type-of-deal of-a-man, with a-
An interesting thing to note about the late-19th century and early-twentieth-century English-language German-speaking literature is its penchant for the Latinate. The latter-day Germanophile is apt to be a bit of a snob, and the Romantic-tinged Germanophile is apt to be a bit of a snob as well (and not in the sense that one would have to be in order to be a proper Anglo-Saxon.
The latter is a bit of a stretch, but the former is a bit more plausible. The point is that the whole business of-a-sort of-German-with-a-
An interesting thing that happened to me came to light as a result of the recent fuss about the re-establishment of the Russian Orthodox Church in the West. The first thing that comes to mind of course is the iconoccending 19th century Russian Orthodox Church of the Russian Empire. But it is the more-or-less orthodox Greek Orthodox Orthodox Church of the 18th century of the Orthodox Church of the 19th century in the West.
The Orthodox Church of the 19th century in the West was the so-called autocephrite (or autoceph-something-to-something-something
An interesting thing about the English-speaking world is that it has a tendency to be rather parochial in its outlook. The British Isles has a rather narrow view of the worlda view that is in some ways a sort of a-historical a-historical, that is, in terms of the term British-ness, in the sense of the more-or-less-non-Anglo-Saxon speaking Anglo-Saxon.
In the case of the United States, it has been a rather long-standing tendency to be rather provincialistic. The British-ness of the American-ness is a rather recent development, and in any case the American-ness
An interesting thing about this book is its origins. The author, Richard Llew-Jones, was a man of letters and a novelist, and a man of letters does not write novels. The fact that a man of letters is a man of letters could be a matter of course, but it has been a matter of course-ness for a century. The only thing about the affair that was a bit of a surprise was the extent to which the two were conjoined.
The author's name was used as a pseudonym for the author of the "The Lord of the Rings" trilogy, which is to say that the author's name was used
An interesting thing about the English language is that it has been so widely used as a lingua franca of sorts for a long time, it's rather surprising how much stuff is in there in the sense that it's English-ish (and I use that term in the broadest sense possible sense sense of the word, which is to say that it's not just English-ish, but rather, a kind of English-ish in the sense that it's-ish, if you follow me-ish, you get the drift? -ishness. The point is that the language is rather, to use the technical term, a-centric
The above example is a more abstract feature. Note that in this case, we can increase top_k higher, without degeneration. Also, notice that even though all query tokens are the same (a hyphen), the steering feature can still be applied.
Finally, letās look at how strength can affect steering.
Steering on Gemma 2 2B using a "Psychosis and Schizophrenia" feature, whose queries are defined below. Use the tabs above to view outputs with different values of strength.
Queries
Let's talk about the future of work and how it will affect your organization. In this session, we look at trends that are changing organizations today through an employee-centric lens to understand what employees want from their employers in a post COVID world
About our Speaker
Anant Agarwal is CEO & Founder @ edX , Board Member@ MITx . An expert on online learning with 20+ years experience leading technology companies including Microsoft Research Labs; Director for Internet Services - Windows Division (1985); Executive Vice President/Chief Technology Officer ā Akamai Technologies Inc.; Chairman Emeritusā University College London Institute For Learning Science( UCLILS)
Let's talk about the 10th anniversary of this beautiful, wonderful series. I know that it was not on my list but if you think carefully and deeply then maybe there is a reason why we are going to discuss today one thing in particular which has been happening since last year or so...
There have always been those who say "I don't like anything from Naruto" (which means Shippuden) because they were too young when all these things happened during their childhood days! It doesn't matter how old she gets; her work will never change for anyone else except herself alone forevermore without any explanation whatsoever either way around!! So let
Let's talk about the topic of today. Today we are going to be talking specifically on how you can get rid of back pain and I know that this is a subject matter where there has been so much misconception, especially in our time now because people don't understand what they should do when it comes to their body issues like for example with regards to neck pains or whatever else but let us just focus upon one aspect at hand which would definitely deal directly into your spine problems here as well if only those who suffer from them could have access towards some kind off relief through any type whatsoever then maybe things might actually work out alright without being too difficult afterwards either way
Let's talk about it.
I was watching the first episode of season 3 and I noticed something that really bothered me, so much in fact...that is why this theory came into my mind....
This thing happened at least three times throughout all episodes:
And then there were these moments which are also characteristic for schizophrenia (but not only):
The following picture shows a part with an obvious schizophrenic-like aura
A lot can be said here but let us come to another subject - depression! The main character has been depressed since childhood because she lost her mother when young girl during some kind of accident . And now we have what you see
Let's talk about. Now, I want to concentrate on the first part of this disorder which is called psychosis and then we will go through it all in detail now because there are some other disorders that exist like mania for example but they donāt have psychotic features associated with them so those ones would be left out here or something else might occur as well ,etc., etc.. But if you look at a person from an outside perspective without any underlying conditions such things usually happen during acute phase when someone gets affected by depression/anxiety anxiety-depression spectrum (ie depressive episodes).
Now let me explain what i mean before going into details . So basically these symptoms
Itās clear that this method of steering isnāt perfect, and in practice we found it about as robust as SAE steering. It is still future work to make it more robust and general - but the good news is that the parameters you can change are much more sensitive than in SAE steering.
We will be releasing a frontend soon, that allows you to test steering yourself. Please stay tuned!
Toxicity Detection
// Coming soon!
Ablations
Figure 6a. Reconstruction error decreases as number of quantizers increase. Experiment done on Gemma 2 9B. As opposed to Figure 3b, this ablation actually trains a different RQAE for each num_quantizer.
// TODO: More ablation studies
// codebook_size_per_dim vs reconstruction
// codebook_size_per_dim vs codebook_dim
Conclusion
RQAE is a new type of model. Weāre not sure if it should replace SAEs, but we think that it clearly has benefits that SAEs donāt have. Since it does have much lower reconstruction error, and can still perform many of the common interpretability tasks that SAEs can perform, we think it is a promising direction for future work.
Weāll be releasing code, models, and a frontend to interact with steering soon. Weāll update this post when that happens. If you want to stay updated, follow me on twitter.