zoeyuchao TongheZhangTH commited on
Commit
a0cbb89
·
verified ·
1 Parent(s): 187a52a

Upload folder using huggingface_hub (#1)

Browse files

- Upload folder using huggingface_hub (dc34d7f6754037362a4772c80fce01cfff94153f)


Co-authored-by: Tonghe-Zhang <TongheZhangTH@users.noreply.huggingface.co>

checkpoints/global_step_150/actor/assets/global_step_150/meta/norm_stats.json ADDED
@@ -0,0 +1,84 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "norm_stats": {
3
+ "state": {
4
+ "mean": [
5
+ 0.07715192926514126,
6
+ 0.17562641970592635,
7
+ 0.3297820340626251,
8
+ -0.00012817750709427617,
9
+ 1.0532821746532337,
10
+ 0.28805883864091536,
11
+ 0.032687542157139704,
12
+ 0.0320219144458349
13
+ ],
14
+ "std": [
15
+ 0.1568252537660898,
16
+ 0.17647293867360334,
17
+ 0.24448783802580087,
18
+ 0.008376862380625637,
19
+ 0.16392024213549403,
20
+ 0.8762590087586914,
21
+ 0.004601313696099623,
22
+ 0.005041351585429002
23
+ ],
24
+ "q01": [
25
+ -0.20252828299999237,
26
+ -0.21026643007993698,
27
+ -0.21924416363239288,
28
+ -0.017773685976862907,
29
+ 0.8226010286808014,
30
+ -1.5134326148033141,
31
+ 0.021859830990433694,
32
+ 0.021078145131468773
33
+ ],
34
+ "q99": [
35
+ 0.41199415922164917,
36
+ 0.48613785028457646,
37
+ 0.7463473677635193,
38
+ 0.016534310169517997,
39
+ 1.4169902801513672,
40
+ 1.8775047111511234,
41
+ 0.03700000047683716,
42
+ 0.03700000047683716
43
+ ]
44
+ },
45
+ "actions": {
46
+ "mean": [
47
+ 0.0002789121082322397,
48
+ 0.0009369332012308343,
49
+ -0.002918429722345843,
50
+ -0.00008228545217631092,
51
+ -0.0011254942718871975,
52
+ -0.005657345530526205,
53
+ 0.4473280228642576
54
+ ],
55
+ "std": [
56
+ 0.007706928797192953,
57
+ 0.008309105883277136,
58
+ 0.01254204095154902,
59
+ 0.0017527195470117924,
60
+ 0.005567940097160219,
61
+ 0.05726387888461874,
62
+ 0.8943707590531353
63
+ ],
64
+ "q01": [
65
+ -0.025044564604759217,
66
+ -0.03014469236135483,
67
+ -0.03136833757162094,
68
+ -0.005981205031275749,
69
+ -0.02122248336672783,
70
+ -0.20235729217529297,
71
+ -1.0
72
+ ],
73
+ "q99": [
74
+ 0.02858110427856446,
75
+ 0.03280403457581998,
76
+ 0.03535533085465432,
77
+ 0.005980716086924076,
78
+ 0.013883259352296595,
79
+ 0.2048933058977127,
80
+ 1.0
81
+ ]
82
+ }
83
+ }
84
+ }
checkpoints/global_step_150/actor/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0adb5b40f064760ff245f53ae0a4e078b86680b65f77d1f5991a173543edd5a1
3
+ size 8537972300
model_config.yaml ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model_name: openpi
2
+ precision: bf16
3
+ num_action_chunks: 5
4
+ action_dim: 7
5
+ is_lora: false
6
+ num_images_in_input: 1
7
+ use_proprio: true
8
+ num_steps: 4
9
+ logprob_gather_mode: sum
10
+ openpi:
11
+ simulator: maniskill
12
+ noise_level: 0.5
13
+ action_chunk: ${actor.model.num_action_chunks}
14
+ num_steps: ${actor.model.num_steps}
15
+ train_expert_only: true
16
+ action_env_dim: ${actor.model.action_dim}
17
+ noise_method: reinflow
18
+ noise_anneal: false
19
+ action_horizon: 8
20
+ noise_params:
21
+ - 0.16
22
+ - 0.12
23
+ - 200
24
+ adv_method: ppo
25
+ pi05: true
26
+ value_after_vlm: true
27
+ joint_logprob: true
28
+ policy_setup: widowx_bridge
29
+ sharding_strategy: full_shard
run_ppo_pi05_maniskill.log ADDED
The diff for this file is too large to render. See raw diff