Skip to content

Commit d198501

Browse files
authored
Release v0.2.1 (#208)
1 parent 8bd9ac3 commit d198501

File tree

5 files changed

+10
-3
lines changed

5 files changed

+10
-3
lines changed

README.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -22,6 +22,8 @@
2222

2323
## 🚀 News
2424

25+
* [2025-08] ✨ Trinity-RFT v0.2.1 is released with enhanced features for Agentic RL and Async RL.
26+
* [2025-08] 🎵 We introduce [CHORD](https://github.com/modelscope/Trinity-RFT/tree/main/examples/mix_chord), a dynamic integration of SFT and RL for enhanced LLM fine-tuning ([paper](https://arxiv.org/pdf/2508.11408)).
2527
* [2025-08] We now support training on general multi-step workflows! Please check out examples for [ALFWorld](./docs/sphinx_doc/source/tutorial/example_step_wise.md) and [ReAct](./docs/sphinx_doc/source/tutorial/example_react.md).
2628
* [2025-07] Trinity-RFT v0.2.0 is released.
2729
* [2025-07] We update the [technical report](https://arxiv.org/abs/2505.17826) (arXiv v2) with new features, examples, and experiments.

README_zh.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -22,6 +22,8 @@
2222

2323
## 🚀 最新动态
2424

25+
* [2025-08] ✨ 发布 Trinity-RFT v0.2.1 版本,强化了 Agentic RL 和 异步 RL 相关功能。
26+
* [2025-08] 🎵 我们推出了 [CHORD](https://github.com/modelscope/Trinity-RFT/tree/main/examples/mix_chord),一种动态整合 SFT 和 RL 来微调 LLM 的方法([论文](https://arxiv.org/pdf/2508.11408))。
2527
* [2025-08] Trinity-RFT 现在已经支持通用多轮工作流的训练了,请参考 [ALFWorld](./docs/sphinx_doc/source/tutorial/example_step_wise.md)[ReAct](./docs/sphinx_doc/source/tutorial/example_react.md) 的例子!
2628
* [2025-07] 发布 Trinity-RFT v0.2.0 版本,新增了多项功能优化。
2729
* [2025-07] 更新了[技术报告](https://arxiv.org/abs/2505.17826) (arXiv v2),增加了新功能、示例和实验。

docs/sphinx_doc/source/main.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,8 @@
88

99
## 🚀 News
1010

11+
* [2025-08] ✨ Trinity-RFT v0.2.1 is released with enhanced features for Agentic RL and Async RL.
12+
* [2025-08] 🎵 We introduce [CHORD](https://github.com/modelscope/Trinity-RFT/tree/main/examples/mix_chord), a dynamic integration of SFT and RL for enhanced LLM fine-tuning ([paper](https://arxiv.org/pdf/2508.11408)).
1113
* [2025-08] We now support training on general multi-step workflows! Please check out examples for [ALFWorld](./docs/sphinx_doc/source/tutorial/example_step_wise.md) and [ReAct](./docs/sphinx_doc/source/tutorial/example_react.md).
1214
* [2025-07] Trinity-RFT v0.2.0 is released.
1315
* [2025-07] We update the [technical report](https://arxiv.org/abs/2505.17826) (arXiv v2) with new features, examples, and experiments.

trinity/trainer/verl/dp_actor.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,6 @@
3535
from trinity.algorithm.kl_fn.kl_fn import DummyKLFn
3636
from trinity.algorithm.utils import prefix_metrics
3737
from trinity.common.config import AlgorithmConfig
38-
from trinity.utils.plugin_loader import load_plugins
3938

4039
__all__ = ["DataParallelPPOActor"]
4140

@@ -49,7 +48,6 @@ def __init__(
4948
):
5049
"""When optimizer is None, it is Reference Policy"""
5150
super().__init__(config, actor_module, actor_optimizer)
52-
load_plugins()
5351
self.policy_loss_fn = None
5452
self.kl_loss_fn = None
5553
self.entropy_loss_fn = None

trinity/utils/plugin_loader.py

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,10 @@ def load_plugin_from_dirs(plugin_dirs: Union[str, List[str]]) -> None:
4646
continue
4747
logger.info(f"Loading plugin modules from [{file}]...")
4848
# load modules from file
49-
load_from_file(os.path.join(plugin_dir, file))
49+
try:
50+
load_from_file(os.path.join(plugin_dir, file))
51+
except Exception as e:
52+
logger.error(f"Failed to load plugin from {file}: {e}")
5053

5154

5255
def load_from_file(file_path: str):

0 commit comments

Comments
 (0)