Upload folder using huggingface_hub
Browse files
.gitattributes
CHANGED
|
@@ -74,3 +74,4 @@ hammer/expert-v2/data/main_data.hdf5 filter=lfs diff=lfs merge=lfs -text
|
|
| 74 |
minigrid/fourrooms-random-v0/data/main_data.hdf5 filter=lfs diff=lfs merge=lfs -text
|
| 75 |
minigrid/fourrooms-v0/data/main_data.hdf5 filter=lfs diff=lfs merge=lfs -text
|
| 76 |
antmaze/large-diverse-v1/data/main_data.hdf5 filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 74 |
minigrid/fourrooms-random-v0/data/main_data.hdf5 filter=lfs diff=lfs merge=lfs -text
|
| 75 |
minigrid/fourrooms-v0/data/main_data.hdf5 filter=lfs diff=lfs merge=lfs -text
|
| 76 |
antmaze/large-diverse-v1/data/main_data.hdf5 filter=lfs diff=lfs merge=lfs -text
|
| 77 |
+
antmaze/large-play-v1/data/main_data.hdf5 filter=lfs diff=lfs merge=lfs -text
|
antmaze/large-play-v1/data/main_data.hdf5
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:21a6b23fa819a04376ad4a88def29331442fc4570dfc7b63168c04a61db6dda1
|
| 3 |
+
size 605199408
|
antmaze/large-play-v1/data/metadata.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"total_episodes": 1000, "total_steps": 1000000, "data_format": "hdf5", "observation_space": "{\"type\": \"Dict\", \"subspaces\": {\"achieved_goal\": {\"type\": \"Box\", \"dtype\": \"float64\", \"shape\": [2], \"low\": [-Infinity, -Infinity], \"high\": [Infinity, Infinity]}, \"desired_goal\": {\"type\": \"Box\", \"dtype\": \"float64\", \"shape\": [2], \"low\": [-Infinity, -Infinity], \"high\": [Infinity, Infinity]}, \"observation\": {\"type\": \"Box\", \"dtype\": \"float64\", \"shape\": [27], \"low\": [-Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity, -Infinity], \"high\": [Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity, Infinity]}}}", "action_space": "{\"type\": \"Box\", \"dtype\": \"float32\", \"shape\": [8], \"low\": [-1.0, -1.0, -1.0, -1.0, -1.0, -1.0, -1.0, -1.0], \"high\": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0]}", "env_spec": "{\"id\": \"AntMaze_Large-v4\", \"entry_point\": \"gymnasium_robotics.envs.maze.ant_maze_v4:AntMazeEnv\", \"reward_threshold\": null, \"nondeterministic\": false, \"max_episode_steps\": 1000, \"order_enforce\": true, \"disable_env_checker\": false, \"kwargs\": {\"maze_map\": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 1], [1, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1], [1, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1], [1, 0, 1, 1, 1, 1, 0, 1, 1, 1, 0, 1], [1, 0, 0, 1, 0, 1, 0, 0, 0, 0, 0, 1], [1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 1], [1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 0, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], \"reward_type\": \"sparse\", \"continuing_task\": true, \"reset_target\": false}, \"additional_wrappers\": [], \"vector_entry_point\": null}", "dataset_id": "D4RL/antmaze/large-play-v1", "algorithm_name": "QIteration+SAC", "author": ["Alex Davey"], "author_email": ["[email protected]"], "code_permalink": "https://github.com/rodrigodelazcano/d4rl-minari-dataset-generation", "minari_version": "0.4.3", "eval_env_spec": "{\"id\": \"AntMaze_Large-v4\", \"entry_point\": \"gymnasium_robotics.envs.maze.ant_maze_v4:AntMazeEnv\", \"reward_threshold\": null, \"nondeterministic\": false, \"max_episode_steps\": 1000, \"order_enforce\": true, \"disable_env_checker\": false, \"kwargs\": {\"maze_map\": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, \"r\", 0, 0, 0, 1, 0, 0, 0, 0, 0, 1], [1, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1], [1, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 1], [1, 0, 1, 1, 1, 1, 0, 1, 1, 1, 0, 1], [1, 0, 0, 1, 0, 1, 0, 0, 0, 0, 0, 1], [1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 1], [1, 0, 0, 1, 0, 0, 0, 1, 0, \"g\", 0, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], \"reward_type\": \"sparse\", \"continuing_task\": true, \"reset_target\": false}, \"additional_wrappers\": [], \"vector_entry_point\": null}", "ref_max_score": 353.5400085449219, "ref_min_score": 0.0, "num_episodes_average_score": 100, "dataset_size": 605.2, "description": "The data is collected from the [`AntMaze_Large-v4`](https://robotics.farama.org/envs/maze/ant_maze/) environment. At the beginning of each episode random locations for the goal and agent's reset are selected. The success rate of all the trajectories is more than 80%, failed trajectories occur because the Ant flips and can't stand up again. Also note that when the Ant reaches the goal the episode doesn't terminate or generate a new target leading to a reward accumulation. The Ant reaches the goals by following a set of waypoints using a goal-reaching policy trained using SAC.", "requirements": ["mujoco>=3.1.1,<=3.1.6", "gymnasium-robotics>=1.2.3"]}
|