Skip to content

Question about input of Mesh Transformer #74

Answered by MarcusLoppe
inmny asked this question in Q&A
Discussion options

You must be logged in to vote

Hey,

So the embedding they are talking about is not a vector embedding but tokens/indices to a codebook.
The output of the encoder is vector embedding but that is then quantized ( as per paper), the output of the quantization is the codes / tokens.

So I think it's maybe the paper authors just use different words to explain the concepts, as far as i know the embedding can be just a number value (token), it's not always the case that it's a vector.

You can think it as the face embedding from the encoder as a vector which then is compressed to a slot in a codebook , if you were to use the 192 vector embedding as "tokens" it would require too much resources and the output of the transformer w…

Replies: 1 comment 2 replies

Comment options

You must be logged in to vote
2 replies
@YoungLee1995
Comment options

@MarcusLoppe
Comment options

Answer selected by inmny
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
3 participants