diff --git a/README.md b/README.md index 3f47a00..741b2d3 100644 --- a/README.md +++ b/README.md @@ -35,7 +35,6 @@ Firstly, each algorithm is implemented strictly according to the original paper - [Overview of Algorithms](#overview-of-algorithms) - [Supported Environments: Safety-Gymnasium](#supported-environments-safety-gymnasium) - - [Gymnasium-based Environments](#gymnasium-based-environments) - [Isaac Gym-based Environments](#isaac-gym-based-environments) - [Selected Tasks](#selected-tasks) - [Pre-requisites](#pre-requisites) @@ -75,7 +74,6 @@ Here we provide a table of Safe RL algorithms that the benchmark includes. For more details, please refer to [Safety-Gymnasium](https://github.com/PKU-Alignment/safety-gymnasium). -### Gymnasium-based Environments @@ -145,7 +143,12 @@ For more details, please refer to [Safety-Gymnasium](https://github.com/PKU-Alig
-**note**: Safe velocity tasks support both single-agent and multi-agent algorithms, while safe navigation tasks only support single-agent algorithms currently. +**note**: + +- **Safe Velocity** and **Safe Isaac Gym** tasks support both single-agent and multi-agent algorithms. +- **Safe Navigation** tasks support single-agent algorithms. +- **Safe MultiGoal** tasks support multi-agent algorithms. +- **Safe Isaac Gym** tasks do not support evaluation after training yet. ### Isaac Gym-based Environments diff --git a/safepo/evaluate.py b/safepo/evaluate.py index dee85cb..2818ea7 100644 --- a/safepo/evaluate.py +++ b/safepo/evaluate.py @@ -137,7 +137,7 @@ def single_runs_eval(eval_dir, eval_episodes): config_path = eval_dir + '/config.json' config = json.load(open(config_path, 'r')) env = config['task'] if 'task' in config.keys() else config['env_name'] - if env in multi_agent_velocity_map.keys(): + if env in multi_agent_velocity_map.keys() or env in multi_agent_goal_tasks: reward, cost = eval_multi_agent(eval_dir, eval_episodes) else: reward, cost = eval_single_agent(eval_dir, eval_episodes)