Skip to content

PCA on what? #100

Answered by Pringled
sachinruk asked this question in Q&A
Discussion options

You must be logged in to vote

Hi @sachinruk, we are applying PCA directly on the static token embeddings that you get by forward passing the vocabulary. The relevant line can be found here. So essentially, we first forward pass all the tokens, which gives you (vocab_size, dim_size) embeddings (where dim_size is the dimensionality of the model you are distilling), and then we apply PCA on those embeddings, which gives you (vocab_size, pca_dims) output embeddings. Hope that answers your question!

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by Pringled
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants