Skip to content

Commit b588c42

Browse files
committed
edit documentation
1 parent 5ff54a9 commit b588c42

File tree

6 files changed

+70
-21
lines changed

6 files changed

+70
-21
lines changed

docs/make.jl

Lines changed: 15 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -12,21 +12,22 @@ makedocs(
1212
##############################################
1313
"Basics" => [
1414
"index.md",
15-
# "install.md",
15+
"mdp.md",
1616
# "get_started.md",
1717
# "concepts.md"
1818
],
19-
"Usage" => [
20-
"mdp.md",
21-
# "pomdp.md",
22-
# "simplegame.md",
23-
# "mg.md",
24-
# "pomg.md",
25-
# "decpomdp.md"
26-
],
19+
# "Usage" => [
20+
# "mdp.md",
21+
# # "pomdp.md",
22+
# # "simplegame.md",
23+
# # "mg.md",
24+
# # "pomg.md",
25+
# # "decpomdp.md"
26+
# ],
2727

2828

2929
"MDP Models" => [
30+
# "mdp.md",
3031
"hexworld.md",
3132
"2048.md",
3233
"cart_pole.md",
@@ -36,26 +37,31 @@ makedocs(
3637
],
3738

3839
"POMDP Models" => [
40+
"pomdp.md",
3941
"crying_baby.md",
4042
"machine_replacement.md",
4143
"catch.md"
4244
],
4345

4446
"Simple Games" => [
47+
"simplegame.md",
4548
"prisoners_dilemma.md",
4649
"rock_paper_scissors.md",
4750
"travelers.md"
4851
],
4952

5053
"POMG Models" => [
54+
"pomg.md",
5155
"multicaregiver.md",
5256
],
5357

5458
"Markov Game" => [
59+
"mg.md",
5560
"predator_prey.md",
5661
],
5762

5863
"Dec-POMDP" => [
64+
"decpomdp.md",
5965
"collab_predator_prey.md",
6066
],
6167
# "Concepts" => [

docs/src/decpomdp.md

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,14 @@
1+
# DecPOMDP Usage
2+
3+
## Decentralized POMDP
4+
The DecPOMDP struct gives the following objects:
5+
- `γ`: discount factor
6+
- ``: agents
7+
- `𝒮`: state space
8+
- `𝒜`: joint action space
9+
- `𝒪`: joint observation space
10+
- `T`: transition function
11+
- `O`: joint observation function
12+
- `R`: joint reward function
13+
14+
The agents `` are the players of the game. The joint action space `𝒜` is the set of all possible ordered pairs of actions amongst all of the agents. The joint observation space `𝒪` is the set of all possible joint observations. The transition function takes in a state `s` in `𝒮`, a joint action `a` and a new state `s'`and returns the transition probability of going from `s` to `s'` by taking action `a`. The joint observation function takes in a state, `s`, a joint action, `a`, and a joint observation `o` in `𝒪` and returns a probability of observing `o` by taking action `a` from state `s`. The joint reward function `R` takes a state and a joint action in `𝒜` and returns a reward value.

docs/src/index.md

Lines changed: 5 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -5,13 +5,6 @@
55

66
![Problem Summary](figures/problemsum.svg)
77

8-
### Usage
9-
10-
```@contents
11-
Pages = [ "mdp.md" ]
12-
```
13-
14-
158
### MDP Models
169

1710
```@contents
@@ -21,29 +14,29 @@ Pages = [ "hexworld.md", "2048.md", "cart_pole.md", "mountain_car.md", "simple_l
2114
### POMDP Models
2215

2316
```@contents
24-
Pages = [ "crying_baby.md", "machine_replacement.md", "catch.md" ]
17+
Pages = [ "pomdp.md", "crying_baby.md", "machine_replacement.md", "catch.md" ]
2518
```
2619

2720
### Simple Games
2821

2922
```@contents
30-
Pages = [ "prisoners_dilemma.md", "rock_paper_scissors.md", "travelers.md" ]
23+
Pages = [ "simplegame.md", "prisoners_dilemma.md", "rock_paper_scissors.md", "travelers.md" ]
3124
```
3225

3326
### POMG Models
3427

3528
```@contents
36-
Pages = [ "multicaregiver.md" ]
29+
Pages = [ "pomg.md", "multicaregiver.md" ]
3730
```
3831

3932
### Markov Games
4033

4134
```@contents
42-
Pages = [ "predator_prey.md" ]
35+
Pages = [ "mg.md", "predator_prey.md" ]
4336
```
4437

4538
### Dec-POMDP
4639

4740
```@contents
48-
Pages = [ "collab_predator_prey.md" ]
41+
Pages = [ "decpomdp.md", "collab_predator_prey.md" ]
4942
```

docs/src/mg.md

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,12 @@
1+
# MG Usage
2+
3+
## Markov Game
4+
The MG struct gives the following objects:
5+
- `γ`: discount factor
6+
- ``: agents
7+
- `𝒮`: state space
8+
- `𝒜`: joint action space
9+
- `T`: transition function
10+
- `R`: joint reward function
11+
12+
The agents `` are the players of the game. The joint action space `𝒜` is the set of all possible ordered pairs of actions amongst all of the agents. The transition function takes in a state `s` in `𝒮`, a joint action `a` and a new state `s'` and returns the transition probability of going from `s` to `s'` by taking action `a`. The joint reward function `R` takes a state and a joint action in `𝒜` and returns a reward value.

docs/src/pomg.md

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,14 @@
1+
# POMG Usage
2+
3+
## Partially Observable Markov Game
4+
The POMG struct gives the following objects:
5+
- `γ`: discount factor
6+
- ``: agents
7+
- `𝒮`: state space
8+
- `𝒜`: joint action space
9+
- `𝒪`: joint observation space
10+
- `T`: transition function
11+
- `O`: joint observation function
12+
- `R`: joint reward function
13+
14+
The agents `` are the players of the game. The joint action space `𝒜` is the set of all possible ordered pairs of actions amongst all of the agents. The joint observation space `𝒪` is the set of all possible joint observations. The transition function takes in a state `s` in `𝒮`, a joint action `a` and a new state `s'`and returns the transition probability of going from `s` to `s'` by taking action `a`. The joint observation function takes in a state, `s`, a joint action, `a`, and a joint observation `o` in `𝒪` and returns a probability of observing `o` by taking action `a` from state `s`. The joint reward function `R` takes a state and a joint action in `𝒜` and returns a reward value.

docs/src/simplegame.md

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,10 @@
1+
# SimpleGame Usage
2+
3+
## Simple Game
4+
The SimpleGame struct gives the following objects:
5+
- `γ`: discount factor
6+
- ``: agents
7+
- `𝒜`: joint action space
8+
- `R`: joint reward function
9+
10+
The agents `` in a simple game are the players of the game. The joint action space `𝒜` is the set of all possible ordered pairs of actions amongst all of the agents. The joint reward function `R` takes a joint action in `𝒜` and returns a reward value.

0 commit comments

Comments
 (0)