Skip to content

Commit 81756b7

Browse files
committed
fix modeing
1 parent bb45cb8 commit 81756b7

File tree

5 files changed

+15
-9
lines changed

5 files changed

+15
-9
lines changed

β€Žpaddlenlp/transformers/__init__.pyβ€Ž

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -215,6 +215,7 @@
215215
from .layoutxlm.modeling import *
216216
from .layoutxlm.tokenizer import *
217217
from .llama import *
218+
from .llm_embed.modeling import *
218219
from .luke.configuration import *
219220
from .luke.modeling import *
220221
from .luke.tokenizer import *
@@ -246,6 +247,7 @@
246247
from .nezha.configuration import *
247248
from .nezha.modeling import *
248249
from .nezha.tokenizer import *
250+
from .nv_embed.modeling import *
249251
from .nystromformer.configuration import *
250252
from .nystromformer.modeling import *
251253
from .nystromformer.tokenizer import *
@@ -320,5 +322,3 @@
320322
from .xlm_roberta.tokenizer import *
321323
from .xlm_roberta.configuration import *
322324
from .yuan import *
323-
from .llm_embed import *
324-
from .nv_embed import *

β€Žpaddlenlp/transformers/llm_embed/modeling.pyβ€Ž

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -34,10 +34,14 @@ class EncoderOutput(ModelOutput):
3434
scores: Optional[paddle.Tensor] = None
3535

3636

37+
__all__ = ["BiEncoderModel"]
38+
39+
3740
class BiEncoderModel(PretrainedModel):
3841
def __init__(
3942
self,
4043
model_name_or_path: str = None,
44+
dtype: str = "float16",
4145
normalized: bool = False,
4246
sentence_pooling_method: str = "cls",
4347
negatives_cross_device: bool = False,
@@ -53,7 +57,7 @@ def __init__(
5357
max_seq_length: int = 4096,
5458
):
5559
super().__init__()
56-
self.model = AutoModel.from_pretrained(model_name_or_path, convert_from_torch=True)
60+
self.model = AutoModel.from_pretrained(model_name_or_path, dtype=dtype, convert_from_torch=True)
5761
self.model_config = AutoConfig.from_pretrained(model_name_or_path)
5862
self.cross_entropy = nn.CrossEntropyLoss(reduction="mean")
5963

β€Žpaddlenlp/transformers/nv_embed/modeling.pyβ€Ž

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -26,6 +26,8 @@
2626
from .. import AutoTokenizer, MistralModel, PretrainedConfig, PretrainedModel
2727
from ..model_outputs import BaseModelOutputWithPast, ModelOutput
2828

29+
__all__ = ["NVEncodeModel"]
30+
2931

3032
@dataclass
3133
class EncoderOutput(ModelOutput):
@@ -121,7 +123,7 @@ def forward(self, last_hidden_states, pool_mask):
121123
one = paddle.eye(
122124
num_rows=self.config.hidden_size,
123125
num_columns=self.config.hidden_size,
124-
dtype=str(self.latents.weight.dtype).split(".")[-1],
126+
dtype=self.latents.weight.dtype,
125127
)
126128
self_latents_weight_T = self.latents(one).T
127129
# latents = repeat(self_latents_weight_T, "d h -> b d h", b=last_hidden_states.shape[0]) # from einops import repeat
@@ -206,7 +208,8 @@ def __init__(
206208
self.latent_model = LatentModel(config=config) # get latent model structure
207209

208210
self.tokenizer = AutoTokenizer.from_pretrained(tokenizer_path, padding_side="right")
209-
self.tokenizer.pad_token = self.tokenizer.eos_token
211+
if self.tokenizer.pad_token is None:
212+
self.tokenizer.pad_token = self.tokenizer.eos_token
210213

211214
self.query_instruction = query_instruction
212215
self.document_instruction = document_instruction
@@ -495,6 +498,7 @@ def encode_sentences(self, sentences: List[str], instruction_len, **kwargs) -> n
495498
sentences_batch,
496499
max_length=4096,
497500
padding=True,
501+
return_attention_mask=True,
498502
return_token_type_ids=False,
499503
return_tensors="pd",
500504
truncation=True,

β€Žtests/transformers/llm_embed/test_modeling.pyβ€Ž

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -19,11 +19,10 @@
1919

2020
from paddlenlp.transformers import AutoTokenizer, BiEncoderModel
2121

22-
from ...testing_utils import require_gpu, slow
22+
from ...testing_utils import require_gpu
2323

2424

2525
class BiEncoderModelIntegrationTest(unittest.TestCase):
26-
@slow
2726
@require_gpu(1)
2827
def test_model_tiny_logits(self):
2928
input_texts = [

β€Žtests/transformers/nv_embed/test_modeling.pyβ€Ž

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -19,11 +19,10 @@
1919

2020
from paddlenlp.transformers import NVEncodeModel, PretrainedConfig
2121

22-
from ...testing_utils import require_gpu, slow
22+
from ...testing_utils import require_gpu
2323

2424

2525
class NVEncodeModelIntegrationTest(unittest.TestCase):
26-
@slow
2726
@require_gpu(1)
2827
def test_model_tiny_logits(self):
2928
input_texts = [

0 commit comments

Comments
Β (0)