Skip to content

Commit 42333d2

Browse files
committed
Update readme
1 parent ec6edd4 commit 42333d2

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -95,7 +95,7 @@ All the agents have been evaluated against each other under controlled condition
9595
- Max 3 sec per move
9696
- 100 games per pairwise evaluation—exactly balanced across the four starting configurations—which color and who starts—except when a human is involved (only 5 games then)
9797
- Original grid (5 x 5)
98-
- Since all the top algorithms (MCTS and Minimax) are deterministic, we need to add small randomness to prevent each starting configuration from leading to the same game. We set `tau=.5, p_mix=0, a_dirichlet=0`.
98+
- Since all the top algorithms (MCTS and Minimax) are deterministic, we need to add small randomness to prevent each starting configuration from leading to the same game. We set `tau=.5, p_mix=0, a_dirichlet=0` to randomize the MCTS agents; the Minimax ones being les performant, we didn't bother with randomizing them for now.
9999
- Hardware information:
100100
- CPU: AMD Ryzen 9 5900HX with Radeon Graphics (8 cores, 16 threads)
101101
- GPU: No NVIDIA GPU
@@ -104,7 +104,7 @@ All the agents have been evaluated against each other under controlled condition
104104
- Python 3.12.9 (CPython)
105105
- Libraries: {'numpy': '2.0.1', 'torch': '2.6.0+cpu'}
106106

107-
See [comparison.ipynb](notebooks/comparison.ipynb) for code reproducibility.
107+
See [benchmark.ipynb](notebooks/benchmark.ipynb) for code reproducibility.
108108

109109
Below is the pairwise algorithm comparison; the value for some row R and column C corresponds to the win rate of R against C. For example, I (human) beat MCTS deep Q-learning 20% of the time.
110110

0 commit comments

Comments
 (0)