Skip to content

Commit d30f507

Browse files
committed
+ fix NomicBertModel
Signed-off-by: wang.yuqi <[email protected]>
1 parent be1d139 commit d30f507

File tree

2 files changed

+2
-3
lines changed

2 files changed

+2
-3
lines changed

vllm/model_executor/models/bert_with_rope.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,6 @@
2222
from vllm.model_executor.layers.vocab_parallel_embedding import (
2323
VocabParallelEmbedding)
2424
from vllm.model_executor.model_loader.weight_utils import default_weight_loader
25-
from vllm.model_executor.models import SupportsV0Only
2625
from vllm.model_executor.models.interfaces import SupportsQuant
2726
from vllm.model_executor.models.utils import WeightsMapper
2827
from vllm.sequence import IntermediateTensors
@@ -466,7 +465,7 @@ def load_weights(self, weights: Iterable[tuple[str,
466465
return loaded_params
467466

468467

469-
class NomicBertModel(BertWithRope, SupportsV0Only):
468+
class NomicBertModel(BertWithRope):
470469
# for https://huggingface.co/nomic-ai/nomic-bert-2048
471470

472471
hf_to_vllm_mapper = WeightsMapper(

vllm/model_executor/models/config.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -93,7 +93,7 @@ def verify_and_update_config(vllm_config: "VllmConfig") -> None:
9393
config.num_hidden_layers = config.n_layer
9494

9595
head_dim = config.hidden_size // config.num_attention_heads
96-
rotary_emb_dim = head_dim * config.rotary_emb_fraction
96+
rotary_emb_dim = int(head_dim * config.rotary_emb_fraction)
9797
max_trained_positions = getattr(config, "max_trained_positions", 2048)
9898
config.rotary_kwargs = {
9999
"head_size": head_dim,

0 commit comments

Comments
 (0)