Commit
·
410a202
1
Parent(s):
73e69d2
Initial commit
Browse files- README.md +1 -1
- args.yml +1 -1
- results.json +1 -1
- tqc-ReachCube-v0.zip +2 -2
- tqc-ReachCube-v0/actor.optimizer.pth +1 -1
- tqc-ReachCube-v0/critic.optimizer.pth +1 -1
- tqc-ReachCube-v0/data +17 -17
- tqc-ReachCube-v0/ent_coef_optimizer.pth +1 -1
- tqc-ReachCube-v0/policy.pth +1 -1
- tqc-ReachCube-v0/pytorch_variables.pth +1 -1
- train_eval_metrics.zip +2 -2
README.md
CHANGED
@@ -16,7 +16,7 @@ model-index:
|
|
16 |
type: ReachCube-v0
|
17 |
metrics:
|
18 |
- type: mean_reward
|
19 |
-
value: -
|
20 |
name: mean_reward
|
21 |
verified: false
|
22 |
---
|
|
|
16 |
type: ReachCube-v0
|
17 |
metrics:
|
18 |
- type: mean_reward
|
19 |
+
value: -59.79 +/- 11.56
|
20 |
name: mean_reward
|
21 |
verified: false
|
22 |
---
|
args.yml
CHANGED
@@ -56,7 +56,7 @@
|
|
56 |
- - save_replay_buffer
|
57 |
- false
|
58 |
- - seed
|
59 |
-
-
|
60 |
- - storage
|
61 |
- null
|
62 |
- - study_name
|
|
|
56 |
- - save_replay_buffer
|
57 |
- false
|
58 |
- - seed
|
59 |
+
- 2637980760
|
60 |
- - storage
|
61 |
- null
|
62 |
- - study_name
|
results.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"mean_reward": -
|
|
|
1 |
+
{"mean_reward": -59.785499800000004, "std_reward": 11.556857543993244, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2024-06-13T15:15:15.088761"}
|
tqc-ReachCube-v0.zip
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:55a5982517a496788579699504d85475d718fba9808b4883fc32228d9a7249f0
|
3 |
+
size 3418285
|
tqc-ReachCube-v0/actor.optimizer.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 589975
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7fd1c103619dd7e7c35d1d9cc69706b5b972ed1fec244cf32b224c38f9e58d37
|
3 |
size 589975
|
tqc-ReachCube-v0/critic.optimizer.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1255594
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:511b1ee5bcff7b14c9803d69f203b748b837d069359021ce426956405afd2ce4
|
3 |
size 1255594
|
tqc-ReachCube-v0/data
CHANGED
@@ -4,20 +4,20 @@
|
|
4 |
":serialized:": "gAWVMQAAAAAAAACMGHNiM19jb250cmliLnRxYy5wb2xpY2llc5SMEE11bHRpSW5wdXRQb2xpY3mUk5Qu",
|
5 |
"__module__": "sb3_contrib.tqc.policies",
|
6 |
"__doc__": "\n Policy class (with both actor and critic) for TQC.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param use_expln: Use ``expln()`` function instead of ``exp()`` when using gSDE to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param clip_mean: Clip the mean output when using gSDE to avoid numerical instability.\n :param features_extractor_class: Features extractor to use.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n :param n_quantiles: Number of quantiles for the critic.\n :param n_critics: Number of critic networks to create.\n :param share_features_extractor: Whether to share or not the features extractor\n between the actor and the critic (this saves computation time)\n ",
|
7 |
-
"__init__": "<function MultiInputPolicy.__init__ at
|
8 |
"__abstractmethods__": "frozenset()",
|
9 |
-
"_abc_impl": "<_abc._abc_data object at
|
10 |
},
|
11 |
"verbose": 1,
|
12 |
"policy_kwargs": {
|
13 |
"use_sde": true
|
14 |
},
|
15 |
-
"num_timesteps":
|
16 |
"_total_timesteps": 10000000,
|
17 |
"_num_timesteps_at_start": 0,
|
18 |
"seed": 0,
|
19 |
"action_noise": null,
|
20 |
-
"start_time":
|
21 |
"learning_rate": 0.0003,
|
22 |
"tensorboard_log": null,
|
23 |
"_last_obs": null,
|
@@ -27,25 +27,25 @@
|
|
27 |
},
|
28 |
"_last_original_obs": {
|
29 |
":type:": "<class 'collections.OrderedDict'>",
|
30 |
-
":serialized:": "
|
31 |
-
"arm_qpos": "[[-
|
32 |
-
"arm_qvel": "[[-
|
33 |
-
"cube_pos": "[[
|
34 |
},
|
35 |
-
"_episode_num":
|
36 |
"use_sde": true,
|
37 |
"sde_sample_freq": -1,
|
38 |
-
"_current_progress_remaining": 0.
|
39 |
"_stats_window_size": 100,
|
40 |
"ep_info_buffer": {
|
41 |
":type:": "<class 'collections.deque'>",
|
42 |
-
":serialized:": "
|
43 |
},
|
44 |
"ep_success_buffer": {
|
45 |
":type:": "<class 'collections.deque'>",
|
46 |
":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="
|
47 |
},
|
48 |
-
"_n_updates":
|
49 |
"observation_space": {
|
50 |
":type:": "<class 'gymnasium.spaces.dict.Dict'>",
|
51 |
":serialized:": "gAWVzAMAAAAAAACMFWd5bW5hc2l1bS5zcGFjZXMuZGljdJSMBERpY3SUk5QpgZR9lCiMBnNwYWNlc5SMC2NvbGxlY3Rpb25zlIwLT3JkZXJlZERpY3SUk5QpUpQojAhhcm1fcXBvc5SMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWBgAAAAAAAAABAQEBAQGUaBOMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLBoWUjAFDlHSUUpSMDWJvdW5kZWRfYWJvdmWUaBwolgYAAAAAAAAAAQEBAQEBlGggSwaFlGgkdJRSlIwGX3NoYXBllEsGhZSMA2xvd5RoHCiWGAAAAAAAAADbD0nA2w9JwNsPScDbD0nA2w9JwNsPScCUaBZLBoWUaCR0lFKUjARoaWdolGgcKJYYAAAAAAAAANsPSUDbD0lA2w9JQNsPSUDbD0lA2w9JQJRoFksGhZRoJHSUUpSMCGxvd19yZXBylIwKLTMuMTQxNTkyN5SMCWhpZ2hfcmVwcpSMCTMuMTQxNTkyN5SMCl9ucF9yYW5kb22UTnVijAhhcm1fcXZlbJRoDSmBlH2UKGgQaBZoGWgcKJYGAAAAAAAAAAEBAQEBAZRoIEsGhZRoJHSUUpRoJ2gcKJYGAAAAAAAAAAEBAQEBAZRoIEsGhZRoJHSUUpRoLEsGhZRoLmgcKJYYAAAAAAAAAAAAIMEAACDBAAAgwQAAIMEAACDBAAAgwZRoFksGhZRoJHSUUpRoM2gcKJYYAAAAAAAAAAAAIEEAACBBAAAgQQAAIEEAACBBAAAgQZRoFksGhZRoJHSUUpRoOIwFLTEwLjCUaDqMBDEwLjCUaDxOdWKMCGN1YmVfcG9zlGgNKYGUfZQoaBBoFmgZaBwolgMAAAAAAAAAAQEBlGggSwOFlGgkdJRSlGgnaBwolgMAAAAAAAAAAQEBlGggSwOFlGgkdJRSlGgsSwOFlGguaBwolgwAAAAAAAAAAAAgwQAAIMEAACDBlGgWSwOFlGgkdJRSlGgzaBwolgwAAAAAAAAAAAAgQQAAIEEAACBBlGgWSwOFlGgkdJRSlGg4jAUtMTAuMJRoOowEMTAuMJRoPE51YnVoLE5oEE5oPE51Yi4=",
|
@@ -83,12 +83,12 @@
|
|
83 |
"__module__": "stable_baselines3.common.buffers",
|
84 |
"__annotations__": "{'observation_space': <class 'gymnasium.spaces.dict.Dict'>, 'obs_shape': typing.Dict[str, typing.Tuple[int, ...]], 'observations': typing.Dict[str, numpy.ndarray], 'next_observations': typing.Dict[str, numpy.ndarray]}",
|
85 |
"__doc__": "\n Dict Replay buffer used in off-policy algorithms like SAC/TD3.\n Extends the ReplayBuffer to use dictionary observations\n\n :param buffer_size: Max number of element in the buffer\n :param observation_space: Observation space\n :param action_space: Action space\n :param device: PyTorch device\n :param n_envs: Number of parallel environments\n :param optimize_memory_usage: Enable a memory efficient variant\n Disabled for now (see https://github.com/DLR-RM/stable-baselines3/pull/243#discussion_r531535702)\n :param handle_timeout_termination: Handle timeout termination (due to timelimit)\n separately and treat the task as infinite horizon task.\n https://github.com/DLR-RM/stable-baselines3/issues/284\n ",
|
86 |
-
"__init__": "<function DictReplayBuffer.__init__ at
|
87 |
-
"add": "<function DictReplayBuffer.add at
|
88 |
-
"sample": "<function DictReplayBuffer.sample at
|
89 |
-
"_get_samples": "<function DictReplayBuffer._get_samples at
|
90 |
"__abstractmethods__": "frozenset()",
|
91 |
-
"_abc_impl": "<_abc._abc_data object at
|
92 |
},
|
93 |
"replay_buffer_kwargs": {},
|
94 |
"train_freq": {
|
|
|
4 |
":serialized:": "gAWVMQAAAAAAAACMGHNiM19jb250cmliLnRxYy5wb2xpY2llc5SMEE11bHRpSW5wdXRQb2xpY3mUk5Qu",
|
5 |
"__module__": "sb3_contrib.tqc.policies",
|
6 |
"__doc__": "\n Policy class (with both actor and critic) for TQC.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param use_expln: Use ``expln()`` function instead of ``exp()`` when using gSDE to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param clip_mean: Clip the mean output when using gSDE to avoid numerical instability.\n :param features_extractor_class: Features extractor to use.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n :param n_quantiles: Number of quantiles for the critic.\n :param n_critics: Number of critic networks to create.\n :param share_features_extractor: Whether to share or not the features extractor\n between the actor and the critic (this saves computation time)\n ",
|
7 |
+
"__init__": "<function MultiInputPolicy.__init__ at 0x7f9ebb4677f0>",
|
8 |
"__abstractmethods__": "frozenset()",
|
9 |
+
"_abc_impl": "<_abc._abc_data object at 0x7f9ebb47c2c0>"
|
10 |
},
|
11 |
"verbose": 1,
|
12 |
"policy_kwargs": {
|
13 |
"use_sde": true
|
14 |
},
|
15 |
+
"num_timesteps": 224928,
|
16 |
"_total_timesteps": 10000000,
|
17 |
"_num_timesteps_at_start": 0,
|
18 |
"seed": 0,
|
19 |
"action_noise": null,
|
20 |
+
"start_time": 1718291358584308985,
|
21 |
"learning_rate": 0.0003,
|
22 |
"tensorboard_log": null,
|
23 |
"_last_obs": null,
|
|
|
27 |
},
|
28 |
"_last_original_obs": {
|
29 |
":type:": "<class 'collections.OrderedDict'>",
|
30 |
+
":serialized:": "gAWVrwQAAAAAAACMC2NvbGxlY3Rpb25zlIwLT3JkZXJlZERpY3SUk5QpUpQojAhhcm1fcXBvc5SMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJaAAQAAAAAAAG4Xb7xkr+c96E8Nv0zbBT9s9vg9tEZ/v3H9jb06M0g9qfPWvhiYKT9kDxs/9E1pvwfXpr7ELFk9HwimvvYbGT9oeuk+LExfv4UFir7CI/c8St7mvqFhND8swMk+zKJBvzg3/b0RmiM9IKm8vpxWVj+kOvM81g1nv0l3D77OYA48ZxbRvjc8tD7eP24+mjhsvzjzibxSyCo9hjHuvsRpKz88k5c+Fhmev/V9JL6Fwga9TccRv2/ZoT5Cez8+6tNzv6QKib10gEK9Iu1avmN53j6vsrw+k0B/v+SNOr36Igy9+tKivp8bKD9jSSg+qOpbvz77Sb0mPqe9mBKvvsAUCz9EHtY+d0WHv4r9nb7zk0m9aaxTvnvLOj/i32S90f2Bv98jQb6hIh8+N1Hxvlv0iD97Ca0+b2prv7mIyL2SNIG8O9mfvmXJOD+RpgG97vKGv74sBL7Y26e9sMXCvpbdGj/GDJ4+QEEfvzGQFr2U90C9TNu7vk8uHT+vdwA+taeQv5SMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJLEEsGhpSMAUOUdJRSlIwIYXJtX3F2ZWyUaAcoloABAAAAAAAAwiLNPlWTGkCoEI/A6hiqwGByk0A6H6NAsphrwK/Aa0CRBEHAWZV3wJf0lMBLKZZAeHMfv4RTQT8HG/G/jEIWQH4UoEAj5Fk+RdVPPWdF3b/xUfu+r9NIwIExpMA5IZxAFaNuP8PSaz8l9fe+Xm9CQMAeocC+jp/A0bCXv4HOpT3YAZq/vEOdwHJdoECu4qLAtFKIvvZDJsDHCifACLmowLB/oEAyz6HAYu7Xv0quI8BY2SO/wFXNQNlUksDkkJZAIoY8P48i+Ltw+NC/AiVGQNGSlsA+LZ7AIOY3QO1uC0CZYY++FfwIQK0tO8Dtq57AtY5HQCwvAUCYgWbAEoiCwJYJoEDrlKPAXH7+vkNpBT9scIG/Mh7JQEdYoUBVjp/ARhKbQAZ4mkBmlPE9P3UNQflUbkB0tZ/Au+BZvq1GJEBEYFHAXP+KwGA5n8C1ZppAgscNP/n1yr9AKE8+TZ0SQWS8gcD0YoZActupPs/NFkBcv8s/fJhJQDqqnMDnA6PAlGgOSxBLBoaUaBJ0lFKUjAhjdWJlX3Bvc5RoByiWwAAAAAAAAAA0Sgc+RBNfPjp6dTwo2J29vYhOPjp6dTzXjpc9K4d5PpHpjDwsZVC8gD0wPjp6dTxwko+9ejPoPTp6dTxKhvq9A9NYPjp6dTydl+89Fv4nPjp6dTyBu5Q9gq73PTp6dTydeqy90dR2Pjp6dTysXQ8+WOwsPjp6dTxCSO29D59RPjp6dTx43Kw8/DpqPuswczzYd1a9/+xXPjp6dTwAlZs9OvbWPbHhbzyvRAS+QPjiPTp6dTy/Uig9laQSPgdhhDyUaA5LEEsDhpRoEnSUUpR1Lg==",
|
31 |
+
"arm_qpos": "[[-0.01459299 0.1131275 -0.5520005 0.52287745 0.12156376 -0.9971726 ]\n [-0.06933106 0.04887698 -0.41982773 0.662477 0.6057036 -0.9113457 ]\n [-0.32585928 0.05302121 -0.3242807 0.5980829 0.456012 -0.87225604]\n [-0.26957336 0.03016842 -0.45091468 0.7046147 0.39404428 -0.75639033]\n [-0.12364048 0.03994185 -0.36847782 0.83725905 0.02969105 -0.90255487]\n [-0.14010347 0.00869007 -0.40837404 0.3520219 0.23266551 -0.9227387 ]\n [-0.01683961 0.04169495 -0.46522158 0.6695826 0.29604518 -1.2351406 ]\n [-0.16063674 -0.03290035 -0.56944734 0.316112 0.18699363 -0.9524523 ]\n [-0.06691483 -0.04748578 -0.21379521 0.43451986 0.36855075 -0.9970791 ]\n [-0.04554547 -0.03421304 -0.31801587 0.65667146 0.16434245 -0.8590493 ]\n [-0.04931187 -0.08166151 -0.34193873 0.54328537 0.41819966 -1.0568074 ]\n [-0.30857497 -0.04921336 -0.20671238 0.72966737 -0.05587757 -1.0155584 ]\n [-0.1886134 0.15540554 -0.47132275 1.0699571 0.33796296 -0.9195928 ]\n [-0.09791703 -0.01577214 -0.3122042 0.72182304 -0.03165299 -1.0542886 ]\n [-0.12907693 -0.08196229 -0.3804145 0.60494363 0.3086912 -0.6220894 ]\n [-0.03675861 -0.04711111 -0.36690748 0.61398786 0.12545656 -1.130118 ]]",
|
32 |
+
"arm_qvel": "[[ 4.0065581e-01 2.4152424e+00 -4.4707832e+00 -5.3155413e+00\n 4.6077118e+00 5.0975618e+00]\n [-3.6811948e+00 3.6836355e+00 -3.0159037e+00 -3.8684905e+00\n -4.6548572e+00 4.6925406e+00]\n [-6.2285566e-01 7.5518060e-01 -1.8836373e+00 2.3478117e+00\n 5.0025015e+00 2.1278434e-01]\n [ 5.0740499e-02 -1.7286805e+00 -4.9085954e-01 -3.1379201e+00\n -5.1310430e+00 4.8790555e+00]\n [ 9.3217593e-01 9.2118472e-01 -4.8429218e-01 3.0380473e+00\n -5.0350037e+00 -4.9861746e+00]\n [-1.1850835e+00 8.0960281e-02 -1.2031813e+00 -4.9145184e+00\n 5.0114069e+00 -5.0901709e+00]\n [-2.6625597e-01 -2.5978980e+00 -2.6100328e+00 -5.2725868e+00\n 5.0155869e+00 -5.0565424e+00]\n [-1.6869624e+00 -2.5575128e+00 -6.4003515e-01 6.4167175e+00\n -4.5728574e+00 4.7051868e+00]\n [ 7.3642170e-01 -7.5724791e-03 -1.6325817e+00 3.0960088e+00\n -4.7054219e+00 -4.9430227e+00]\n [ 2.8734207e+00 2.1786454e+00 -2.8004149e-01 2.1403859e+00\n -2.9246628e+00 -4.9584870e+00]\n [ 3.1180851e+00 2.0185041e+00 -3.6016598e+00 -4.0791101e+00\n 5.0011702e+00 -5.1119285e+00]\n [-4.9705780e-01 5.2113742e-01 -1.0112433e+00 6.2849360e+00\n 5.0420260e+00 -4.9861245e+00]\n [ 4.8459806e+00 4.8271513e+00 1.1795883e-01 8.8411245e+00\n 3.7239363e+00 -4.9909000e+00]\n [-2.1277134e-01 2.5668137e+00 -3.2715006e+00 -4.3436718e+00\n -4.9757538e+00 4.8250375e+00]\n [ 5.5382550e-01 -1.5856315e+00 2.0230198e-01 9.1634035e+00\n -4.0542469e+00 4.1995792e+00]\n [ 3.3175236e-01 2.3563116e+00 1.5917773e+00 3.1499319e+00\n -4.8957796e+00 -5.0942264e+00]]",
|
33 |
+
"cube_pos": "[[ 0.132119 0.21784693 0.01498275]\n [-0.07707244 0.20169349 0.01498275]\n [ 0.07400291 0.24367969 0.01720122]\n [-0.01271943 0.1721096 0.01498275]\n [-0.07010353 0.11337943 0.01498275]\n [-0.12232645 0.21174245 0.01498275]\n [ 0.1169884 0.1640552 0.01498275]\n [ 0.07262326 0.12093832 0.01498275]\n [-0.08421824 0.2410462 0.01498275]\n [ 0.14000577 0.16887033 0.01498275]\n [-0.11586048 0.20470832 0.01498275]\n [ 0.02110122 0.22874063 0.01484321]\n [-0.05236039 0.210865 0.01498275]\n [ 0.07596779 0.10496183 0.01464121]\n [-0.12916826 0.11082506 0.01498275]\n [ 0.04109454 0.14320596 0.01615955]]"
|
34 |
},
|
35 |
+
"_episode_num": 448,
|
36 |
"use_sde": true,
|
37 |
"sde_sample_freq": -1,
|
38 |
+
"_current_progress_remaining": 0.9775088,
|
39 |
"_stats_window_size": 100,
|
40 |
"ep_info_buffer": {
|
41 |
":type:": "<class 'collections.deque'>",
|
42 |
+
":serialized:": "gAWVRAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHwE/yX7+DOC6MAWyUTfQBjAF0lEdAa3FAJLM9sHV9lChoBkfAS9N4u9OARWgHTfQBaAhHQGttp5mh/RV1fZQoaAZHwE9WxtYSxqxoB030AWgIR0BraguwosqbdX2UKGgGR8BURp2Qnx8VaAdN9AFoCEdAa2Vwc5sCT3V9lChoBkfARhba0x/NJWgHTfQBaAhHQGzBrS/j81p1fZQoaAZHwFBRihFmWdFoB030AWgIR0BsveqBEroXdX2UKGgGR8BUSIvFm4AkaAdN9AFoCEdAbLpGoaUA1nV9lChoBkfATs25J9RaYGgHTfQBaAhHQGy2rThHbyp1fZQoaAZHwHI6fFaSs8xoB030AWgIR0Bssw/qxC6ZdX2UKGgGR8BR7R20Re1KaAdN9AFoCEdAbK9yy2QXAXV9lChoBkfASpbcCYCyQmgHTfQBaAhHQGyr3mFJxvN1fZQoaAZHwFSHwZOzpotoB030AWgIR0BsqClP8AJcdX2UKGgGR8BI4InKGL1maAdN9AFoCEdAbKSPjn3cpXV9lChoBkfAVIK9sabWmWgHTfQBaAhHQGyg8c+7lJZ1fZQoaAZHwFDBQO4G2ThoB030AWgIR0BsnU+kgwGodX2UKGgGR8BQ/lnuiN83aAdN9AFoCEdAbJmtknTiKnV9lChoBkfAUtgVN5+pfmgHTfQBaAhHQGyWDnmq5sl1fZQoaAZHwFhpvES/TLJoB030AWgIR0BsknXK8tf5dX2UKGgGR8BPmukk8ifQaAdN9AFoCEdAbI7Z+QU5/HV9lChoBkfAViRysCDEnGgHTfQBaAhHQGyKPt2LYPJ1fZQoaAZHwFVVvXsgMc9oB030AWgIR0Bt5pEpiI+GdX2UKGgGR8BJsRptaY/naAdN9AFoCEdAbeLOXVsk6nV9lChoBkfAT7+kDZDiO2gHTfQBaAhHQG3fKk2xY7t1fZQoaAZHwFoZdxQzk6toB030AWgIR0Bt25Cx/ustdX2UKGgGR8BFFu09hZyNaAdN9AFoCEdAbdfzDn/1hHV9lChoBkfAUvilN1yNoGgHTfQBaAhHQG3UVbA1vVF1fZQoaAZHwFzYCROk+HJoB030AWgIR0Bt0MEovzvrdX2UKGgGR8BW1X1anrIHaAdN9AFoCEdAbc0L2pQ1rXV9lChoBkfAUCXgsK9f1GgHTfQBaAhHQG3JccdYGMZ1fZQoaAZHwFILJ0GNaQpoB030AWgIR0BtxdOymhugdX2UKGgGR8BSPljurp7kaAdN9AFoCEdAbcIxQBPsRnV9lChoBkfASwy1LJ0W/WgHTfQBaAhHQG2+juSfUWl1fZQoaAZHwFSJg2Ifr8loB030AWgIR0Btuu9i+cpcdX2UKGgGR8BR57coH9m6aAdN9AFoCEdAbbdWcSXdCXV9lChoBkfAVDNmSQo1DWgHTfQBaAhHQG2zuiWVu791fZQoaAZHwFOl5Jsfq5doB030AWgIR0Btrx6+nIhhdX2UKGgGR8BbAQ04zabnaAdN9AFoCEdAb58HDaXa8HV9lChoBkfAUTu1twaR6mgHTfQBaAhHQG+bRBVuJk51fZQoaAZHwFNBfVqesgdoB030AWgIR0Bvl6AjIJZ4dX2UKGgGR8BREWlImPYGaAdN9AFoCEdAb5QGnn+yaHV9lChoBkfAUgFw0fozN2gHTfQBaAhHQG+QaTGHYYl1fZQoaAZHwFlyklNUOutoB030AWgIR0BvjMvwmVqvdX2UKGgGR8BKO7qyGBWgaAdN9AFoCEdAb4k3gDRtxnV9lChoBkfAU6wVzp5eJGgHTfQBaAhHQG+Fgjps41h1fZQoaAZHwFdxTn7pFCtoB030AWgIR0BvgegvlEJCdX2UKGgGR8BRVEOEug6EaAdN9AFoCEdAb35KLbYbsHV9lChoBkfASid6zE74jGgHTfQBaAhHQG96p/gBLf11fZQoaAZHwE048zyjHn5oB030AWgIR0BvdwWrOqvNdX2UKGgGR8BRAer+5vtMaAdN9AFoCEdAb3NmUW2w3nV9lChoBkfATchNCZ4Oc2gHTfQBaAhHQG9vzW5H3Dh1fZQoaAZHwFWRZa3Zwn9oB030AWgIR0BvbDE3sHB2dX2UKGgGR8BILKlHjIaMaAdN9AFoCEdAb2eV3Ux20XV9lChoBkfAUSl80DU3GWgHTfQBaAhHQHBXeaa1Cw91fZQoaAZHwEkwB2fTTfBoB030AWgIR0BwVZgx8D0UdX2UKGgGR8BQFdipeeFtaAdN9AFoCEdAcFPGMXJo03V9lChoBkfASrxw4sEq2GgHTfQBaAhHQHBR+XRgJC11fZQoaAZHwE0s3xWkrPNoB030AWgIR0BwUCqwQlKLdX2UKGgGR8BJLIfr8iwCaAdN9AFoCEdAcE5cDbJwKnV9lChoBkfATtQL5RCQcWgHTfQBaAhHQHBMkdV/+bV1fZQoaAZHwFl/BHTZxrBoB030AWgIR0BwSrcynDR/dX2UKGgGR8BNzm/336AOaAdN9AFoCEdAcEjqM3qA0HV9lChoBkfASD8CT2WY4WgHTfQBaAhHQHBHGzKLbYd1fZQoaAZHwEzwxbB42TBoB030AWgIR0BwRUoc7yQQdX2UKGgGR8BKNJxFRYRvaAdN9AFoCEdAcEN5AQg9vHV9lChoBkfAXbEdvKlpGmgHTfQBaAhHQHBBqVD8cdZ1fZQoaAZHwEtIoDxLCepoB030AWgIR0BwP9zfaYeDdX2UKGgGR8BUsWxY7q6faAdN9AFoCEdAcD4Ox0MgEHV9lChoBkfATbapPykKu2gHTfQBaAhHQHA7wSFoL5R1fZQoaAZHwE/fpA2Q4jtoB030AWgIR0Bw4xznzQNTdX2UKGgGR8BQGTodMj/uaAdN9AFoCEdAcOE7YkE9uHV9lChoBkfAS9w4bS7XhGgHTfQBaAhHQHDfaVUuL751fZQoaAZHwEfK+PBBRhtoB030AWgIR0Bw3Zx82JizdX2UKGgGR8BjA3MEA5q/aAdN9AFoCEdAcNvNo8IRiHV9lChoBkfAUwUKb8WKuWgHTfQBaAhHQHDZ/ugHu7Z1fZQoaAZHwEdGIAOrhitoB030AWgIR0Bw2DSVnmJWdX2UKGgGR8BNS4593KSxaAdN9AFoCEdAcNZZ5zHS4XV9lChoBkfAZFMn0kGA1GgHTfQBaAhHQHDUjNliBoV1fZQoaAZHwEZ3gc94eLhoB030AWgIR0Bw0r3Gn4widX2UKGgGR8BJ0MTN+so2aAdN9AFoCEdAcNDsguAZsXV9lChoBkfAUONUYKpkw2gHTfQBaAhHQHDPG0/nnuB1fZQoaAZHwEsOq7ROUMZoB030AWgIR0BwzUuL74zrdX2UKGgGR8Bb7cMuvlltaAdN9AFoCEdAcMt/A0sOG3V9lChoBkfAUBl6Y3Ns32gHTfQBaAhHQHDJsPWhAW11fZQoaAZHwFRJ3B55Z8toB030AWgIR0Bwx2M2m52AdX2UKGgGR8BkoMglnh86aAdN9AFoCEdAcXGyad+Xq3V9lChoBkfATdjjPv8ZUGgHTfQBaAhHQHFv0ZR8+id1fZQoaAZHwEk5H1e0G/xoB030AWgIR0Bxbf/vOQhfdX2UKGgGR8BcpnNxEORUaAdN9AFoCEdAcWwzcAR02nV9lChoBkfATFLeIl+mWWgHTfQBaAhHQHFqZPVNHpd1fZQoaAZHwEihqj8DSw5oB030AWgIR0BxaJaOgg5jdX2UKGgGR8BQaeCsfaHsaAdN9AFoCEdAcWbMn7YTTXV9lChoBkfAU6mg3974SGgHTfQBaAhHQHFk8j/uLJl1fZQoaAZHwE3jaCcwxnFoB030AWgIR0BxYyWBz3h5dX2UKGgGR8BFF7el9BrvaAdN9AFoCEdAcWFWweNkv3V9lChoBkfAUEdpFkQPJGgHTfQBaAhHQHFfhdIGyHF1fZQoaAZHwE/HzErGza9oB030AWgIR0BxXbUPQOWjdX2UKGgGR8BHLdG7SRbKaAdN9AFoCEdAcVvlYlpoK3V9lChoBkfARmLKA8Swn2gHTfQBaAhHQHFaGOlwcYJ1fZQoaAZHwFDlw1R+BpZoB030AWgIR0BxWErNGEwndX2UKGgGR8BOOFhgE2YOaAdN9AFoCEdAcVX9IPK+z3VlLg=="
|
43 |
},
|
44 |
"ep_success_buffer": {
|
45 |
":type:": "<class 'collections.deque'>",
|
46 |
":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="
|
47 |
},
|
48 |
+
"_n_updates": 14051,
|
49 |
"observation_space": {
|
50 |
":type:": "<class 'gymnasium.spaces.dict.Dict'>",
|
51 |
":serialized:": "gAWVzAMAAAAAAACMFWd5bW5hc2l1bS5zcGFjZXMuZGljdJSMBERpY3SUk5QpgZR9lCiMBnNwYWNlc5SMC2NvbGxlY3Rpb25zlIwLT3JkZXJlZERpY3SUk5QpUpQojAhhcm1fcXBvc5SMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWBgAAAAAAAAABAQEBAQGUaBOMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLBoWUjAFDlHSUUpSMDWJvdW5kZWRfYWJvdmWUaBwolgYAAAAAAAAAAQEBAQEBlGggSwaFlGgkdJRSlIwGX3NoYXBllEsGhZSMA2xvd5RoHCiWGAAAAAAAAADbD0nA2w9JwNsPScDbD0nA2w9JwNsPScCUaBZLBoWUaCR0lFKUjARoaWdolGgcKJYYAAAAAAAAANsPSUDbD0lA2w9JQNsPSUDbD0lA2w9JQJRoFksGhZRoJHSUUpSMCGxvd19yZXBylIwKLTMuMTQxNTkyN5SMCWhpZ2hfcmVwcpSMCTMuMTQxNTkyN5SMCl9ucF9yYW5kb22UTnVijAhhcm1fcXZlbJRoDSmBlH2UKGgQaBZoGWgcKJYGAAAAAAAAAAEBAQEBAZRoIEsGhZRoJHSUUpRoJ2gcKJYGAAAAAAAAAAEBAQEBAZRoIEsGhZRoJHSUUpRoLEsGhZRoLmgcKJYYAAAAAAAAAAAAIMEAACDBAAAgwQAAIMEAACDBAAAgwZRoFksGhZRoJHSUUpRoM2gcKJYYAAAAAAAAAAAAIEEAACBBAAAgQQAAIEEAACBBAAAgQZRoFksGhZRoJHSUUpRoOIwFLTEwLjCUaDqMBDEwLjCUaDxOdWKMCGN1YmVfcG9zlGgNKYGUfZQoaBBoFmgZaBwolgMAAAAAAAAAAQEBlGggSwOFlGgkdJRSlGgnaBwolgMAAAAAAAAAAQEBlGggSwOFlGgkdJRSlGgsSwOFlGguaBwolgwAAAAAAAAAAAAgwQAAIMEAACDBlGgWSwOFlGgkdJRSlGgzaBwolgwAAAAAAAAAAAAgQQAAIEEAACBBlGgWSwOFlGgkdJRSlGg4jAUtMTAuMJRoOowEMTAuMJRoPE51YnVoLE5oEE5oPE51Yi4=",
|
|
|
83 |
"__module__": "stable_baselines3.common.buffers",
|
84 |
"__annotations__": "{'observation_space': <class 'gymnasium.spaces.dict.Dict'>, 'obs_shape': typing.Dict[str, typing.Tuple[int, ...]], 'observations': typing.Dict[str, numpy.ndarray], 'next_observations': typing.Dict[str, numpy.ndarray]}",
|
85 |
"__doc__": "\n Dict Replay buffer used in off-policy algorithms like SAC/TD3.\n Extends the ReplayBuffer to use dictionary observations\n\n :param buffer_size: Max number of element in the buffer\n :param observation_space: Observation space\n :param action_space: Action space\n :param device: PyTorch device\n :param n_envs: Number of parallel environments\n :param optimize_memory_usage: Enable a memory efficient variant\n Disabled for now (see https://github.com/DLR-RM/stable-baselines3/pull/243#discussion_r531535702)\n :param handle_timeout_termination: Handle timeout termination (due to timelimit)\n separately and treat the task as infinite horizon task.\n https://github.com/DLR-RM/stable-baselines3/issues/284\n ",
|
86 |
+
"__init__": "<function DictReplayBuffer.__init__ at 0x7f9ebbe1e3b0>",
|
87 |
+
"add": "<function DictReplayBuffer.add at 0x7f9ebbe1e440>",
|
88 |
+
"sample": "<function DictReplayBuffer.sample at 0x7f9ebbe1e4d0>",
|
89 |
+
"_get_samples": "<function DictReplayBuffer._get_samples at 0x7f9ebbe1e560>",
|
90 |
"__abstractmethods__": "frozenset()",
|
91 |
+
"_abc_impl": "<_abc._abc_data object at 0x7f9ebbe18f00>"
|
92 |
},
|
93 |
"replay_buffer_kwargs": {},
|
94 |
"train_freq": {
|
tqc-ReachCube-v0/ent_coef_optimizer.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1940
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1cb3541c1439c347c73829cec54d9581a32e778652a9eb0fd9095c05e346be8c
|
3 |
size 1940
|
tqc-ReachCube-v0/policy.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1548985
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a32f920bc47b7d526bd3da96b44856834a01934636dd1c28dd4ca57cdb4f22aa
|
3 |
size 1548985
|
tqc-ReachCube-v0/pytorch_variables.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1180
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6dd5f9297ac7540d06679aef37b1afc2737a61e0514cdc96e918437e4bdcb20a
|
3 |
size 1180
|
train_eval_metrics.zip
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7f8fe15969785aafe678ebe5748b0b531c0935bfcdaa0a7dfc021c29228a38c3
|
3 |
+
size 19046
|