Skip to content

Commit ab5f6e7

Browse files
committed
style(nyz): polish flake8 style
1 parent 04a586b commit ab5f6e7

File tree

3 files changed

+4
-3
lines changed

3 files changed

+4
-3
lines changed

ding/reward_model/__init__.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,4 +15,4 @@
1515
from .icm_reward_model import ICMRewardModel
1616
# LLM/VLM reward model and verifier
1717
from .math_reward_model import MathRewardModel
18-
from .math_rule_reward_model import MathRuleRewardModel
18+
from .math_rule_reward_model import MathRuleRewardModel

ding/reward_model/math_reward_model.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -42,4 +42,4 @@ def collect_data(self, data: list) -> None:
4242
pass
4343

4444
def clear_data(self) -> None:
45-
pass
45+
pass

ding/reward_model/math_rule_reward_model.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,8 @@ def estimate(self, data: List[str]) -> List[Dict]:
4141
"""
4242
# 1. parse the query to get question and predicted answer
4343
# 2. get the ground truth answer according to the question
44-
# 3. calculate the reward based on the predicted answer and the ground truth answer (format error -2, answer error -1, correct 1)
44+
# 3. calculate the reward based on the predicted answer and the ground truth answer
45+
# (format error -2, answer error -1, correct 1)
4546
pass
4647

4748
# rule-based reward model does not need training, thus the following methods are empty

0 commit comments

Comments
 (0)