diff --git a/vllm/model_executor/models/mixtral.py b/vllm/model_executor/models/mixtral.py index 6cb1d84965ecf..b68e525d365f4 100644 --- a/vllm/model_executor/models/mixtral.py +++ b/vllm/model_executor/models/mixtral.py @@ -293,7 +293,7 @@ def __init__( self.embed_tokens = VocabParallelEmbedding( config.vocab_size, config.hidden_size, - org_num_embeddings=self.org_vocab_size, + org_num_embeddings=self.vocab_size, ) self.layers = nn.ModuleList([ MixtralDecoderLayer(config, linear_method=linear_method)