Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feat: sebulba ff_ippo #1088

Open
wants to merge 134 commits into
base: develop
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from 86 commits
Commits
Show all changes
134 commits
Select commit Hold shift + click to select a range
adc2114
feat: gym wrapper
Louay-Ben-nessir Jun 10, 2024
ce86d09
chore : pre-commit hooks
Louay-Ben-nessir Jun 10, 2024
d5edf45
fix: merged the observations and action mask
Louay-Ben-nessir Jun 14, 2024
f891be5
fix: Create the gym wrappers directly
Louay-Ben-nessir Jun 14, 2024
15f4867
chore: pre-commit
Louay-Ben-nessir Jun 14, 2024
82ea827
fix: fixed the async env creation
Louay-Ben-nessir Jun 14, 2024
4e94df5
fix: gymV26 compatability wrapper
Louay-Ben-nessir Jun 14, 2024
8a86be9
fix: various minor fixes
Louay-Ben-nessir Jun 15, 2024
1da5c15
fix: handling rware reset function
Louay-Ben-nessir Jun 15, 2024
4466044
feat: async env wrapper , changed the gym wrapper to rware wrapper
Louay-Ben-nessir Jun 16, 2024
24d8aae
fix: fixed the async env wrapper
Louay-Ben-nessir Jun 16, 2024
a6deae2
fix: info only contains the action_mask and reformated (n_agents, n_e…
Louay-Ben-nessir Jun 18, 2024
1475bd0
chore: removed async gym wrapper
Louay-Ben-nessir Jun 22, 2024
9fce9c6
feat: gym metric tracker wrapper
Louay-Ben-nessir Jun 22, 2024
055a326
feat: init sebulba ippo
Louay-Ben-nessir Jun 10, 2024
a435a0a
feat: initial learner / training loop
Louay-Ben-nessir Jun 13, 2024
7e80d7b
fix: changes the env creation
Louay-Ben-nessir Jun 14, 2024
b961336
fix: fixed function calls
Louay-Ben-nessir Jun 15, 2024
502730d
fix: fixed the training and added training logger
Louay-Ben-nessir Jun 22, 2024
1985729
fix: changed the anakin ppo type import
Louay-Ben-nessir Jun 22, 2024
89ed246
feat: fulll sebulba functional
Louay-Ben-nessir Jun 25, 2024
7f43a33
fix: logging and added LBF
Louay-Ben-nessir Jul 2, 2024
8a87258
fix: batch size calc for multiple devices
Louay-Ben-nessir Jul 4, 2024
7f0acd9
fix: num_updates and code refactoring
Louay-Ben-nessir Jul 5, 2024
3e352cf
chore : code cleanup + comments + added checkpoint save
Louay-Ben-nessir Jul 8, 2024
bcdaa38
feat: mappo + removed sebulba specifique types and made the rware wra…
Louay-Ben-nessir Jul 8, 2024
7044fbe
fix: removed the sebulba spesifique types
Louay-Ben-nessir Jul 8, 2024
9433f2e
feat: ff_mappo and rec_ippo in sebulba
Louay-Ben-nessir Jul 10, 2024
627215d
fix: removed the lbf import/wrapper
Louay-Ben-nessir Jul 10, 2024
c3b405d
chore: clean up & updated the code to match the sebulba-ff-ippo branch
Louay-Ben-nessir Jul 10, 2024
e40c5d4
chore : pre-commits and some comments
Louay-Ben-nessir Jul 10, 2024
4b17c15
chore: removed unused config file
Louay-Ben-nessir Jul 10, 2024
9ec6b16
feat: sebulba ff_ippo
Louay-Ben-nessir Jul 10, 2024
e5dd71b
chore: pre-commits
Louay-Ben-nessir Jul 10, 2024
af24082
fix: fix the num_updates_in_eval in the last eval
Louay-Ben-nessir Jul 13, 2024
32ac389
fix: fixed the num evals cacls
Louay-Ben-nessir Jul 16, 2024
45ca587
chore : pre commit
Louay-Ben-nessir Jul 16, 2024
d694498
chore: created the anakin and sebulba folders
Louay-Ben-nessir Jul 16, 2024
cb8111f
fix: imports and config paths in systems
Louay-Ben-nessir Jul 16, 2024
d842375
fix: allow for reproducibility
Louay-Ben-nessir Jul 16, 2024
0a1ffd0
chore: pre-commits
Louay-Ben-nessir Jul 16, 2024
f1adc31
chore: pre-commits
Louay-Ben-nessir Jul 16, 2024
3850591
feat: LBF and reproducibility
Louay-Ben-nessir Jul 16, 2024
0a2ee08
feat : lbf
Louay-Ben-nessir Jul 16, 2024
dc92065
fix: sync neptune logging for sebulba to avoid stalling
Louay-Ben-nessir Jul 17, 2024
133a250
fix: added missing lbf import
Louay-Ben-nessir Jul 17, 2024
b938c83
fix: seeds need to python arrays not np arrays
Louay-Ben-nessir Jul 17, 2024
a368476
fix: config and imports for anakin q_learning and sac
Louay-Ben-nessir Jul 17, 2024
32433ff
chore: arch_name for anakin
Louay-Ben-nessir Jul 17, 2024
a68c8e9
fix: sum the rewards when using a shared reward
Louay-Ben-nessir Jul 17, 2024
8cee7ac
fix: configs revamp
Louay-Ben-nessir Jul 17, 2024
e199f3a
chore: pre-commits
Louay-Ben-nessir Jul 17, 2024
2b71d3b
fix: more config changes
Louay-Ben-nessir Jul 17, 2024
e87ad28
chore: pre-commits
Louay-Ben-nessir Jul 17, 2024
2b587c0
chore: renamed arch_name to architecture_name
Louay-Ben-nessir Jul 18, 2024
5ad4d2f
chore: config files rename
Louay-Ben-nessir Jul 18, 2024
432071e
fix; moved from gym to gymnasium
Louay-Ben-nessir Jul 18, 2024
77e6e12
feat: generic gym wrapper
Louay-Ben-nessir Jul 18, 2024
43511fd
feat: using gymnasium async worker
Louay-Ben-nessir Jul 18, 2024
eaf9a1c
chore: pre-commits and annotaions
Louay-Ben-nessir Jul 18, 2024
16c0ac3
fix: config file fixes
Louay-Ben-nessir Jul 18, 2024
18b928d
fix: rware import
Louay-Ben-nessir Jul 18, 2024
19a7765
fix: better agent ids wrapper?
Louay-Ben-nessir Jul 18, 2024
c4a05d6
chore: bunch of minor changes
Louay-Ben-nessir Jul 18, 2024
5595818
chore : annotation
Louay-Ben-nessir Jul 18, 2024
29b1303
chore: comments
Louay-Ben-nessir Jul 19, 2024
669dfbd
feat: restructured the folders
Louay-Ben-nessir Jul 19, 2024
d1f8364
update the gym wrappers
Louay-Ben-nessir Jul 19, 2024
dc641c6
folder re-structuring
Louay-Ben-nessir Jul 19, 2024
0881d2f
fix: removed deprecated jax call
Louay-Ben-nessir Jul 19, 2024
b60cefe
fix: env wrappers fix
Louay-Ben-nessir Jul 19, 2024
21aafbf
fix: config changes
Louay-Ben-nessir Jul 19, 2024
e09fd60
chore: pre-commits
Louay-Ben-nessir Jul 19, 2024
2a6452d
fix: config file fixes
Louay-Ben-nessir Jul 19, 2024
e2f36f9
fix: LBF import
Louay-Ben-nessir Jul 19, 2024
29396c9
fix: Async worker auto-resetting
Louay-Ben-nessir Jul 19, 2024
6de0b1e
chore: minor changes
Louay-Ben-nessir Jul 19, 2024
7584ce5
fixed: annotations and add agent id spaces
Louay-Ben-nessir Jul 22, 2024
e638e9f
fix: fixed the logging deadlock for sebulba
Louay-Ben-nessir Jul 22, 2024
81b0a89
Merge pull request #4 from Louay-Ben-nessir/feat-sebulba-gym-wrapper
Louay-Ben-nessir Jul 22, 2024
0860518
Merge pull request #1090 from Louay-Ben-nessir/chore--anakin-and-sebu…
sash-a Jul 23, 2024
4c0acdc
Merge remote-tracking branch 'upstream/develop' into chore--sebulba-a…
Louay-Ben-nessir Jul 23, 2024
a85aa2f
chore: pre-commits
Louay-Ben-nessir Jul 23, 2024
e504b47
pre-commit
Louay-Ben-nessir Jul 23, 2024
6a1fad4
Merge pull request #1094 from Louay-Ben-nessir/chore--sebulba-arch-up…
OmaymaMahjoub Jul 23, 2024
0cae539
Merge remote-tracking branch 'upstream/feat/sebulba_arch' into seb-ff…
Louay-Ben-nessir Jul 23, 2024
a19056b
feat : major code restructer, non-blocking evalutors
Louay-Ben-nessir Jul 25, 2024
fc80b91
chore: code cleanup and sps calcs and learner threads
Louay-Ben-nessir Jul 26, 2024
18ec08f
feat: shared time steps checker
Louay-Ben-nessir Jul 29, 2024
38e7229
chore: removed unused eval type
Louay-Ben-nessir Jul 29, 2024
5a5e542
chore: config file changes
Louay-Ben-nessir Jul 29, 2024
dcff2a1
fix: fixed stalling at the end of training
Louay-Ben-nessir Jul 29, 2024
d926c54
chore: code cleanup
Louay-Ben-nessir Jul 29, 2024
7e4698a
chore : various changes
Louay-Ben-nessir Jul 29, 2024
6dac8c3
fix: prevent the pipeline from stalling and a lot of cleanup
Louay-Ben-nessir Jul 30, 2024
23b582c
chore : better error messeages
Louay-Ben-nessir Jul 30, 2024
c71dad8
fix: changed the timestep discount
Louay-Ben-nessir Jul 30, 2024
bfea3aa
chore: very nitpicky clean ups
sash-a Jul 30, 2024
de92f5a
feat: pass timestep instead of obs and done and fix potential race co…
sash-a Jul 30, 2024
1465133
fix: deadlock in pipeline
sash-a Jul 30, 2024
6689c49
fix: wasting samples
Louay-Ben-nessir Aug 11, 2024
c506da3
chore: loss unpacking
Louay-Ben-nessir Aug 11, 2024
b24ac34
fix: updated to work with the latest gymnasium
Louay-Ben-nessir Oct 10, 2024
1dfb241
fix: jumanji
Louay-Ben-nessir Oct 10, 2024
fd8aece
fix: removed depricated gymnasium import
Louay-Ben-nessir Oct 10, 2024
ae53415
feat: minor refactor to sebulba utils
sash-a Oct 10, 2024
724d2dc
chore: a few minor changes to code style
sash-a Oct 10, 2024
fa8a996
Merge branch 'develop' into feat/sebulba_arch
sash-a Oct 11, 2024
0a36fdf
Merge branch 'feat/sebulba_arch' into seb-ff-ippo-only
sash-a Oct 11, 2024
47b8e03
fix: update configs to match latest mava
sash-a Oct 11, 2024
8be8037
fix: reshape with multiple learners and system name
sash-a Oct 11, 2024
4748636
fix: safer pipeline.clear()
sash-a Oct 11, 2024
5593bde
feat: avoid unecessary host-device transfers
sash-a Oct 14, 2024
133ea1a
chore: remove some more device transfers
sash-a Oct 14, 2024
9260e9b
chore: better graceful exit
sash-a Oct 14, 2024
d61dcfb
fix: create envs in main thread to avoid deadlocks
sash-a Oct 15, 2024
105d796
chore: use orginal rware and lbf
Louay-Ben-nessir Oct 15, 2024
f292bf3
fix: possible off by one fix
sash-a Oct 16, 2024
d42d732
fix: change to using gym.make to create envs and fix StepType
sash-a Oct 16, 2024
d4359c1
feat: learner env accumulation
Louay-Ben-nessir Oct 17, 2024
7c78478
feat: jit evaluation on cpu
sash-a Oct 17, 2024
aa49c6f
Merge branch 'seb-ff-ippo-only' of github.com:Louay-Ben-nessir/Mava i…
sash-a Oct 17, 2024
c252ffe
fix: timestep calculation with accumulation
Louay-Ben-nessir Oct 17, 2024
fd7a025
feat: shardmap almost working
sash-a Oct 17, 2024
4013a22
feat: shard_map working
sash-a Oct 18, 2024
0e559d9
fix: key use in actor loss
sash-a Oct 19, 2024
0a6bd49
fix: align gym config with other configs
sash-a Oct 19, 2024
641a548
feat: better env creation and safer sharding
sash-a Oct 19, 2024
c0c88bc
chore: minor env typing fixes
sash-a Oct 19, 2024
354159a
Merge branch 'develop' into seb-ff-ippo-only
sash-a Oct 19, 2024
6b2d01c
fix: start actors simultaneously to avoid deadlocks
Louay-Ben-nessir Oct 21, 2024
a13ab65
feat: support for smac
Louay-Ben-nessir Oct 23, 2024
bc55375
chore: pre-commits
Louay-Ben-nessir Oct 23, 2024
c6d460f
fix: random segfault
Louay-Ben-nessir Oct 27, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions mava/configs/arch/anakin.yaml
Original file line number Diff line number Diff line change
@@ -1,4 +1,5 @@
# --- Anakin config ---
architecture_name: anakin

# --- Training ---
num_envs: 16 # Number of vectorised environments per device.
Expand Down
19 changes: 19 additions & 0 deletions mava/configs/arch/sebulba.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,19 @@
# --- Sebulba config ---
architecture_name: sebulba

# --- Training ---
num_envs: 32 # number of environments per thread.

# --- Evaluation ---
evaluation_greedy: False # Evaluate the policy greedily. If True the policy will select
# an action which corresponds to the greatest logit. If false, the policy will sample
# from the logits.
num_eval_episodes: 32 # Number of episodes to evaluate per evaluation.
num_evaluation: 200 # Number of evenly spaced evaluations to perform during training.
absolute_metric: True # Whether the absolute metric should be computed. For more details
# on the absolute metric please see: https://arxiv.org/abs/2209.10485

# --- Sebulba devices config ---
n_threads_per_executor: 1 # num of different threads/env batches per actor
executor_device_ids: [0] # ids of actor devices
learner_device_ids: [0] # ids of learner devices
2 changes: 1 addition & 1 deletion mava/configs/default_ff_ippo.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,5 +3,5 @@ defaults:
- arch: anakin
- system: ppo/ff_ippo
- network: mlp
- env: rware
- env: rware_gym
Louay-Ben-nessir marked this conversation as resolved.
Show resolved Hide resolved
- _self_
7 changes: 7 additions & 0 deletions mava/configs/default_ff_ippo_seb.yaml
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can we call it sebulba instead of seb 🙏

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm fine with both, but wouldn't it be better to rename all files by system_name_arch.yaml, e.g. default_ff_ippo*.yaml becomes ff_ippo_anakin.yaml and ff_ippo_sebulba.yaml? 🙄
Adding the default_ is optional, but it makes the file name too long

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Agreed with removing default, maybe we should change this in another PR though?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sure agree 🙌

Original file line number Diff line number Diff line change
@@ -0,0 +1,7 @@
defaults:
- logger: ff_ippo
- arch: sebulba
- system: ppo/ff_ippo
- network: mlp
- env: rware_gym
- _self_
19 changes: 19 additions & 0 deletions mava/configs/env/lbf_gym.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,19 @@
# ---Environment Configs---
defaults:
- _self_
- scenario: gym-lbf-2s-8x8-2p-2f-coop # [gym-lbf-2s-8x8-2p-2f-coop, gym-lbf-8x8-2p-2f-coop, gym-lbf-2s-10x10-3p-3f, gym-lbf-10x10-3p-3f, gym-lbf-15x15-3p-5f, gym-lbf-15x15-4p-3f, gym-lbf-15x15-4p-5f]

env_name: LevelBasedForaging # Used for logging purposes.

# Defines the metric that will be used to evaluate the performance of the agent.
# This metric is returned at the end of an experiment and can be used for hyperparameter tuning.
eval_metric: episode_return

# Whether the add agents IDs to the observations returned by the environment.
add_agent_id : False

# Whether or not to log the winrate of this environment.
log_win_rate: False

# Weather or not to sum the returned rewards over all of the agents.
use_shared_rewards: True
19 changes: 19 additions & 0 deletions mava/configs/env/rware_gym.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,19 @@
# ---Environment Configs---
defaults:
- _self_
- scenario: gym-rware-tiny-2ag # [gym-rware-tiny-2ag, gym-rware-tiny-4ag, gym-rware-tiny-4ag-easy, gym-rware-small-4ag]

env_name: RobotWarehouse # Used for logging purposes.

# Defines the metric that will be used to evaluate the performance of the agent.
# This metric is returned at the end of an experiment and can be used for hyperparameter tuning.
eval_metric: episode_return

# Whether the add agents IDs to the observations returned by the environment.
add_agent_id : False

# Whether or not to log the winrate of this environment.
log_win_rate: False

# Weather or not to sum the returned rewards over all of the agents.
use_shared_rewards: True
18 changes: 18 additions & 0 deletions mava/configs/env/scenario/gym-lbf-10x10-3p-3f.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
# The config of the 10x10-3p-3f scenario with the VectorObserver set as default
name: LevelBasedForaging
task_name: 10x10-3p-3f

task_config:
field_size: [10,10]
sight: 10
players: 3
max_num_food: 3
max_player_level: 2
force_coop: False
max_episode_steps: 100
min_player_level : 1
min_food_level : null
max_food_level : null

env_kwargs:
{} # there are no scenario specific env_kwargs for this env
18 changes: 18 additions & 0 deletions mava/configs/env/scenario/gym-lbf-15x15-3p-5f.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
# The config of the 15x15-3p-5f scenario with the VectorObserver set as default
name: LevelBasedForaging
task_name: 15x15-3p-5f

task_config:
field_size: [15, 15]
sight: 15
players: 3
max_num_food: 5
max_player_level: 2
force_coop: False
max_episode_steps: 100
min_player_level : 1
min_food_level : null
max_food_level : null

env_kwargs:
{} # there are no scenario specific env_kwargs for this env
18 changes: 18 additions & 0 deletions mava/configs/env/scenario/gym-lbf-15x15-4p-3f.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
# The config of the 15x15-4p-3f scenario with the VectorObserver set as default
name: LevelBasedForaging
task_name: 15x15-4p-3f

task_config:
field_size: [15, 15]
sight: 15
players: 4
max_num_food: 3
max_player_level: 2
force_coop: False
max_episode_steps: 100
min_player_level : 1
min_food_level : null
max_food_level : null

env_kwargs:
{} # there are no scenario specific env_kwargs for this env
18 changes: 18 additions & 0 deletions mava/configs/env/scenario/gym-lbf-15x15-4p-5f.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
# The config of the 15x15-4p-5f scenario with the VectorObserver set as default
name: LevelBasedForaging
task_name: 15x15-4p-5f

task_config:
field_size: [15, 15]
sight: 15
players: 4
max_num_food: 5
max_player_level: 2
force_coop: False
max_episode_steps: 100
min_player_level : 1
min_food_level : null
max_food_level : null

env_kwargs:
{} # there are no scenario specific env_kwargs for this env
18 changes: 18 additions & 0 deletions mava/configs/env/scenario/gym-lbf-2s-10x10-3p-3f.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
# The config of the 2s10x10-3p-3f scenario with the VectorObserver set as default
name: LevelBasedForaging
task_name: 2s-10x10-3p-3f

task_config:
field_size: [10, 10]
sight: 2
players: 3
max_num_food: 3
max_player_level: 2
force_coop: False
max_episode_steps: 100
min_player_level : 1
min_food_level : null
max_food_level : null

env_kwargs:
{} # there are no scenario specific env_kwargs for this env
18 changes: 18 additions & 0 deletions mava/configs/env/scenario/gym-lbf-2s-8x8-2p-2f-coop.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
# The config of the 2s-8x8-2p-2f-coop scenario with the VectorObserver set as default.
name: LevelBasedForaging
task_name: 2s-8x8-2p-2f-coop

task_config:
field_size: [8, 8] # size of the grid to generate.
sight: 2 # field of view of an agent.
players: 2 # number of agents on the grid.
max_num_food: 2 # number of food in the environment.
max_player_level: 2 # maximum level of the agents (inclusive).
force_coop: True # force cooperation between agents.
max_episode_steps: 100 # max number of steps per episode.
min_player_level : 1 # minimum level of the agents (inclusive).
min_food_level : null
max_food_level : null

env_kwargs:
{} # there are no scenario specific env_kwargs for this env
18 changes: 18 additions & 0 deletions mava/configs/env/scenario/gym-lbf-8x8-2p-2f-coop.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
# The config of the 8x8-2p-2f-coop scenario with the VectorObserver set as default
name: LevelBasedForaging
task_name: 8x8-2p-2f-coop

task_config:
field_size: [8, 8]
sight: 8
players: 2
max_num_food: 2
max_player_level: 2
force_coop: True
max_episode_steps: 100
min_player_level : 1
min_food_level : null
max_food_level : null

env_kwargs:
{} # there are no scenario specific env_kwargs for this env
18 changes: 18 additions & 0 deletions mava/configs/env/scenario/gym-rware-small-4ag.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
# The config of the small-4ag environment
name: RobotWarehouse
task_name: small-4ag

task_config:
column_height: 8
shelf_rows: 2
shelf_columns: 3
n_agents: 4
sensor_range: 1
request_queue_size: 4
msg_bits : 0
max_inactivity_steps : null
max_steps : 500
reward_type : 0

env_kwargs:
{} # there are no scenario specific env_kwargs for this env
18 changes: 18 additions & 0 deletions mava/configs/env/scenario/gym-rware-tiny-2ag.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
# The config of the tiny-2ag environment
name: RobotWarehouse
task_name: tiny-2ag

task_config:
column_height: 8
shelf_rows: 1
shelf_columns: 3
n_agents: 2
sensor_range: 1
request_queue_size: 2
msg_bits : 0
max_inactivity_steps : null
max_steps : 500
reward_type : 0

env_kwargs:
{} # there are no scenario specific env_kwargs for this env
18 changes: 18 additions & 0 deletions mava/configs/env/scenario/gym-rware-tiny-4ag-easy.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
# The config of the tiny-4ag-easy environment
name: RobotWarehouse
task_name: tiny-4ag-easy

task_config:
column_height: 8
shelf_rows: 1
shelf_columns: 3
n_agents: 4
sensor_range: 1
request_queue_size: 8
msg_bits : 0
max_inactivity_steps : null
max_steps : 500
reward_type : 0

env_kwargs:
{} # there are no scenario specific env_kwargs for this env
18 changes: 18 additions & 0 deletions mava/configs/env/scenario/gym-rware-tiny-4ag.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,18 @@
# The config of the tiny_4ag environment
name: RobotWarehouse
task_name: tiny-4ag

task_config:
column_height: 8
shelf_rows: 1
shelf_columns: 3
n_agents: 4
sensor_range: 1
request_queue_size: 4
msg_bits : 0
max_inactivity_steps : null
max_steps : 500
reward_type : 0

env_kwargs:
{} # there are no scenario specific env_kwargs for this env
Loading
Loading