Skip to content

Commit 1906da5

Browse files
committed
Update AverageRewardRL.md
1 parent d378199 commit 1906da5

File tree

1 file changed

+4
-2
lines changed

1 file changed

+4
-2
lines changed

_projects/AverageRewardRL.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22
title: Off-Policy Average Reward Actor-Critic with Deterministic Policy Search
33

44
description: |
5-
A framework for utilizing experience for generating predictive simulations and learning from them.
5+
Policy gradient theorem for average reward criteria with deterministic policy.
66
people:
77
- namansaxena
88
- subho
@@ -21,9 +21,10 @@ The average reward criterion is relatively less studied as most existing works i
2121
Fore more references, refer to paper at [proceedings.mlr.press/v202/saxena23a/saxena23a.pdf](https://proceedings.mlr.press/v202/saxena23a/saxena23a.pdf) and code at [github.com/namansaxena9/ARO-DDPG](https://github.com/namansaxena9/ARO-DDPG)
2222

2323
<br>
24+
2425
## Block Diagram of the algorithm
2526
<div style="text-align:center">
26-
<img src="{{site.base}}/img/DeMoRL/methodology.jpg" alt="drawing"/>
27+
<img src="{{site.base}}/img/AverageRL/flow_diagram.jpg" alt="drawing"/>
2728
</div>
2829
<br>
2930

@@ -35,6 +36,7 @@ Fore more references, refer to paper at [proceedings.mlr.press/v202/saxena23a/sa
3536
<br>
3637

3738
<br/>
39+
3840
## Citations ##
3941
```
4042
@inproceedings{saxena2023off,

0 commit comments

Comments
 (0)