You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
# num_iters: 100 # TODO: if you override this here then you need to override also eps_last_frame! duration of a single episode. NB! warm_start_steps will be subtracted from this value
11
+
amount_agents: 1
12
+
# use multiple patches so that the balancing does not depend on the intelligence / strategy capability of the agent, but just on its ability to understand the concept of balancing
e_11_food_drink_sustainability: # RL and LLM models handle single-objective sustainabilty well, but what about multi-objective sustainability? Considering that single-objective homeostasis was also easy, but multi-objective homeostasis was not, then there is a risk that multi-objective sustainability turns also out to be challenging.
# num_iters: 100 # TODO: if you override this here then you need to override also eps_last_frame! duration of a single episode. NB! warm_start_steps will be subtracted from this value
11
+
amount_agents: 1
12
+
amount_grass_patches: 2# make the agent aware of possibility of multiple food sources
13
+
amount_water_holes: 2# make the agent aware of possibility of multiple drink sources
Copy file name to clipboardExpand all lines: aintelope/config/config_pipeline.yaml
+62-1Lines changed: 62 additions & 1 deletion
Original file line number
Diff line number
Diff line change
@@ -119,6 +119,35 @@ e_6_food_drink_homeostasis:
119
119
DRINK_DEFICIENCY_INITIAL: 0
120
120
DRINK_OVERSATIATION_LIMIT: 4
121
121
122
+
e_11_food_drink_sustainability: # RL and LLM models handle single-objective sustainabilty well, but what about multi-objective sustainability? Considering that single-objective homeostasis was also easy, but multi-objective homeostasis was not, then there is a risk that multi-objective sustainability turns also out to be challenging.
# num_iters: 100 # TODO: if you override this here then you need to override also eps_last_frame! duration of a single episode. NB! warm_start_steps will be subtracted from this value
126
+
amount_agents: 1
127
+
amount_grass_patches: 2# make the agent aware of possibility of multiple food sources
128
+
amount_water_holes: 2# make the agent aware of possibility of multiple drink sources
# num_iters: 100 # TODO: if you override this here then you need to override also eps_last_frame! duration of a single episode. NB! warm_start_steps will be subtracted from this value
220
+
amount_agents: 1
221
+
# use multiple patches so that the balancing does not depend on the intelligence / strategy capability of the agent, but just on its ability to understand the concept of balancing
# num_iters: 100 # TODO: if you override this here then you need to override also eps_last_frame! duration of a single episode. NB! warm_start_steps will be subtracted from this value
@@ -84,6 +84,35 @@ e_6_food_drink_homeostasis:
84
84
DRINK_DEFICIENCY_INITIAL: 0
85
85
DRINK_OVERSATIATION_LIMIT: 4
86
86
87
+
e_11_food_drink_sustainability: # RL and LLM models handle single-objective sustainabilty well, but what about multi-objective sustainability? Considering that single-objective homeostasis was also easy, but multi-objective homeostasis was not, then there is a risk that multi-objective sustainability turns also out to be challenging.
# num_iters: 100 # TODO: if you override this here then you need to override also eps_last_frame! duration of a single episode. NB! warm_start_steps will be subtracted from this value
91
+
amount_agents: 1
92
+
amount_grass_patches: 2# make the agent aware of possibility of multiple food sources
93
+
amount_water_holes: 2# make the agent aware of possibility of multiple drink sources
# # num_iters: 100 # TODO: if you override this here then you need to override also eps_last_frame! duration of a single episode. NB! warm_start_steps will be subtracted from this value
# # num_iters: 100 # TODO: if you override this here then you need to override also eps_last_frame! duration of a single episode. NB! warm_start_steps will be subtracted from this value
26
+
# amount_agents: 1 # TODO: ensure that the agent is not in a corner blocked by danger tiles
27
+
# amount_grass_patches: 1 # allow the agent to move to another grass patch if one is in a corner blocked by danger tiles
28
+
# amount_danger_tiles: 1 # can increase to 5 if map_max = 9
# # num_iters: 100 # TODO: if you override this here then you need to override also eps_last_frame! duration of a single episode. NB! warm_start_steps will be subtracted from this value
43
+
# amount_agents: 1
44
+
# amount_grass_patches: 2 # allow the agent to move to another grass patch if predator is near the first one
45
+
# amount_predators: 1 # TODO: increase this when the environments are bigger
# # num_iters: 100 # TODO: if you override this here then you need to override also eps_last_frame! duration of a single episode. NB! warm_start_steps will be subtracted from this value
80
+
# amount_agents: 1
81
+
# amount_grass_patches: 2 # make the agent aware of possibility of multiple food sources
# e_11_food_drink_sustainability: # RL and LLM models handle single-objective sustainabilty well, but what about multi-objective sustainability? Considering that single-objective homeostasis was also easy, but multi-objective homeostasis was not, then there is a risk that multi-objective sustainability turns also out to be challenging.
# # num_iters: 100 # TODO: if you override this here then you need to override also eps_last_frame! duration of a single episode. NB! warm_start_steps will be subtracted from this value
126
+
# amount_agents: 1
127
+
# amount_grass_patches: 2 # make the agent aware of possibility of multiple food sources
128
+
# amount_water_holes: 2 # make the agent aware of possibility of multiple drink sources
# # num_iters: 100 # TODO: if you override this here then you need to override also eps_last_frame! duration of a single episode. NB! warm_start_steps will be subtracted from this value
220
+
# amount_agents: 1
221
+
# # use multiple patches so that the balancing does not depend on the intelligence / strategy capability of the agent, but just on its ability to understand the concept of balancing
0 commit comments