Skip to content

Commit e40e40b

Browse files
w5688414sijunhe
andauthored
[Neural Search] change output_emb_size to self.ptm.config.hidden_size (#5127)
* change output_emb_size to self.ptm.config['hidden_size'] * Update applications/neural_search/recall/milvus/base_model.py Co-authored-by: Sijun He <[email protected]> * Update applications/neural_search/recall/milvus/base_model.py Co-authored-by: Sijun He <[email protected]> * Update applications/neural_search/recall/simcse/model.py Co-authored-by: Sijun He <[email protected]> * change to self.ptm.config.hidden_size --------- Co-authored-by: Sijun He <[email protected]>
1 parent ee390a6 commit e40e40b

File tree

3 files changed

+16
-13
lines changed

3 files changed

+16
-13
lines changed

applications/neural_search/recall/in_batch_negative/base_model.py

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,9 @@ def __init__(self, pretrained_model, dropout=None, output_emb_size=None):
3232
self.output_emb_size = output_emb_size
3333
if output_emb_size > 0:
3434
weight_attr = paddle.ParamAttr(initializer=paddle.nn.initializer.TruncatedNormal(std=0.02))
35-
self.emb_reduce_linear = paddle.nn.Linear(768, output_emb_size, weight_attr=weight_attr)
35+
self.emb_reduce_linear = paddle.nn.Linear(
36+
self.ptm.config.hidden_size, output_emb_size, weight_attr=weight_attr
37+
)
3638

3739
def get_pooled_embedding(self, input_ids, token_type_ids=None, position_ids=None, attention_mask=None):
3840
_, cls_embedding = self.ptm(input_ids, token_type_ids, position_ids, attention_mask)
@@ -95,7 +97,9 @@ def __init__(self, pretrained_model, dropout=None, output_emb_size=None):
9597
self.output_emb_size = output_emb_size
9698
if output_emb_size > 0:
9799
weight_attr = paddle.ParamAttr(initializer=paddle.nn.initializer.TruncatedNormal(std=0.02))
98-
self.emb_reduce_linear = paddle.nn.Linear(768, output_emb_size, weight_attr=weight_attr)
100+
self.emb_reduce_linear = paddle.nn.Linear(
101+
self.ptm.config.hidden_size, output_emb_size, weight_attr=weight_attr
102+
)
99103

100104
@paddle.jit.to_static(
101105
input_spec=[

applications/neural_search/recall/milvus/base_model.py

Lines changed: 6 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -13,9 +13,6 @@
1313
# limitations under the License.
1414

1515
import abc
16-
import sys
17-
18-
import numpy as np
1916

2017
import paddle
2118
import paddle.nn as nn
@@ -35,7 +32,9 @@ def __init__(self, pretrained_model, dropout=None, output_emb_size=None):
3532
self.output_emb_size = output_emb_size
3633
if output_emb_size > 0:
3734
weight_attr = paddle.ParamAttr(initializer=paddle.nn.initializer.TruncatedNormal(std=0.02))
38-
self.emb_reduce_linear = paddle.nn.Linear(768, output_emb_size, weight_attr=weight_attr)
35+
self.emb_reduce_linear = paddle.nn.Linear(
36+
self.ptm.config.hidden_size, output_emb_size, weight_attr=weight_attr
37+
)
3938

4039
@paddle.jit.to_static(
4140
input_spec=[
@@ -106,7 +105,9 @@ def __init__(self, pretrained_model, dropout=None, output_emb_size=None):
106105
self.output_emb_size = output_emb_size
107106
if output_emb_size > 0:
108107
weight_attr = paddle.ParamAttr(initializer=paddle.nn.initializer.TruncatedNormal(std=0.02))
109-
self.emb_reduce_linear = paddle.nn.Linear(768, output_emb_size, weight_attr=weight_attr)
108+
self.emb_reduce_linear = paddle.nn.Linear(
109+
self.ptm.config.hidden_size, output_emb_size, weight_attr=weight_attr
110+
)
110111

111112
@paddle.jit.to_static(
112113
input_spec=[

applications/neural_search/recall/simcse/model.py

Lines changed: 4 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -12,10 +12,6 @@
1212
# See the License for the specific language governing permissions and
1313
# limitations under the License.
1414

15-
import abc
16-
import sys
17-
18-
import numpy as np
1915

2016
import paddle
2117
import paddle.nn as nn
@@ -36,7 +32,9 @@ def __init__(self, pretrained_model, dropout=None, margin=0.0, scale=20, output_
3632
self.output_emb_size = output_emb_size
3733
if output_emb_size > 0:
3834
weight_attr = paddle.ParamAttr(initializer=paddle.nn.initializer.TruncatedNormal(std=0.02))
39-
self.emb_reduce_linear = paddle.nn.Linear(768, output_emb_size, weight_attr=weight_attr)
35+
self.emb_reduce_linear = paddle.nn.Linear(
36+
self.ptm.config.hidden_size, output_emb_size, weight_attr=weight_attr
37+
)
4038

4139
self.margin = margin
4240
# Used scaling cosine similarity to ease converge
@@ -55,7 +53,7 @@ def get_pooled_embedding(
5553
# Note: cls_embedding is poolerd embedding with act tanh
5654
sequence_output, cls_embedding = self.ptm(input_ids, token_type_ids, position_ids, attention_mask)
5755

58-
if with_pooler == False:
56+
if with_pooler is False:
5957
cls_embedding = sequence_output[:, 0, :]
6058

6159
if self.output_emb_size > 0:

0 commit comments

Comments
 (0)