Skip to content

Commit d49b0d9

Browse files
committed
word2vec demo loss
1 parent c1d6ddd commit d49b0d9

File tree

4 files changed

+11
-8
lines changed

4 files changed

+11
-8
lines changed

models/rank/flen/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -71,7 +71,7 @@ os : windows/linux/macos
7171
python -u ../../../tools/trainer.py -m config.yaml # 全量数据运行config_bigdata.yaml
7272
# 动态图预测
7373
python -u ../../../tools/infer.py -m config.yaml # 全量数据运行config_bigdata.yaml
74-
74+
```
7575

7676
## 模型组网
7777

models/recall/word2vec/dygraph_model.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -85,7 +85,6 @@ def create_metrics(self):
8585
# construct train forward phase
8686
def train_forward(self, dy_model, metrics_list, batch_data, config):
8787
input_word, true_word, neg_word = self.create_feeds(batch_data, config)
88-
8988
true_logits, neg_logits = dy_model.forward(
9089
[input_word, true_word, neg_word])
9190
loss = self.create_loss(true_logits, neg_logits, config)

models/recall/word2vec/static_model.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -95,6 +95,7 @@ def net(self, inputs, is_infer=False):
9595

9696
def create_optimizer(self, strategy=None):
9797
optimizer = paddle.optimizer.SGD(learning_rate=self.learning_rate)
98+
optimizer.minimize(self._cost)
9899
# learning_rate=paddle.fluid.layers.exponential_decay(
99100
# learning_rate=self.learning_rate,
100101
# decay_steps=self.decay_steps,

models/recall/word2vec/word2vec_reader.py

Lines changed: 9 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@
1616
import numpy as np
1717
import io
1818
import six
19-
19+
import time
2020
from paddle.io import IterableDataset
2121

2222

@@ -35,7 +35,7 @@ def __call__(self):
3535
self.idx = 0
3636

3737
result = self.buffer[self.idx]
38-
self.idx += 1
38+
self.idx = self.idx + 1
3939
return result
4040

4141

@@ -52,7 +52,9 @@ def init(self):
5252
self.neg_num = self.config.get("hyper_parameters.neg_num")
5353
self.with_shuffle_batch = self.config.get(
5454
"hyper_parameters.with_shuffle_batch")
55-
self.random_generator = NumpyRandomInt(1, self.window_size + 1)
55+
#self.random_generator = NumpyRandomInt(1, self.window_size + 1)
56+
np.random.seed(12345)
57+
self.random_generator = np.random.randint(1, self.window_size + 1)
5658
self.batch_size = self.config.get("runner.batch_size")
5759

5860
self.cs = None
@@ -78,7 +80,7 @@ def get_context_words(self, words, idx):
7880
idx: input word index
7981
window_size: window size
8082
"""
81-
target_window = self.random_generator()
83+
target_window = self.random_generator
8284
# if (idx - target_window) > 0 else 0
8385
start_point = idx - target_window
8486
if start_point < 0:
@@ -103,8 +105,9 @@ def __iter__(self):
103105
output.append(
104106
np.array([int(context_id)]).astype('int64'))
105107
np.random.seed(12345)
106-
neg_array = self.cs.searchsorted(
107-
np.random.sample(self.neg_num))
108+
tmp = np.random.sample(self.neg_num)
109+
time.sleep(0.0001)
110+
neg_array = self.cs.searchsorted(tmp)
108111
output.append(
109112
np.array([int(str(i))
110113
for i in neg_array]).astype('int64'))

0 commit comments

Comments
 (0)