crazyhathello commited on
Commit
268b296
·
verified ·
1 Parent(s): eb34faf

Upload folder using huggingface_hub

Browse files
Files changed (4) hide show
  1. README.md +9 -0
  2. config.json +61 -0
  3. config.yaml +174 -0
  4. model.safetensors +3 -0
README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - model_hub_mixin
4
+ - pytorch_model_hub_mixin
5
+ ---
6
+
7
+ This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
+ - Library: [More Information Needed]
9
+ - Docs: [More Information Needed]
config.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "beta_end": 0.02,
3
+ "beta_schedule": "squaredcos_cap_v2",
4
+ "beta_start": 0.0001,
5
+ "clip_sample": true,
6
+ "clip_sample_range": 1.0,
7
+ "crop_is_random": false,
8
+ "crop_shape": [
9
+ 320,
10
+ 320
11
+ ],
12
+ "diffusion_step_embed_dim": 128,
13
+ "do_mask_loss_for_padding": false,
14
+ "down_dims": [
15
+ 512,
16
+ 1024,
17
+ 2048
18
+ ],
19
+ "horizon": 16,
20
+ "input_normalization_modes": {
21
+ "observation.images.gripper": "mean_std",
22
+ "observation.images.head": "mean_std",
23
+ "observation.state": "min_max"
24
+ },
25
+ "input_shapes": {
26
+ "observation.images.gripper": [
27
+ 3,
28
+ 320,
29
+ 320
30
+ ],
31
+ "observation.images.head": [
32
+ 3,
33
+ 320,
34
+ 320
35
+ ],
36
+ "observation.state": [
37
+ 9
38
+ ]
39
+ },
40
+ "kernel_size": 5,
41
+ "n_action_steps": 8,
42
+ "n_groups": 8,
43
+ "n_obs_steps": 2,
44
+ "noise_scheduler_type": "DDPM",
45
+ "num_inference_steps": 100,
46
+ "num_train_timesteps": 100,
47
+ "output_normalization_modes": {
48
+ "action": "min_max"
49
+ },
50
+ "output_shapes": {
51
+ "action": [
52
+ 9
53
+ ]
54
+ },
55
+ "prediction_type": "epsilon",
56
+ "pretrained_backbone_weights": null,
57
+ "spatial_softmax_num_keypoints": 32,
58
+ "use_film_scale_modulation": true,
59
+ "use_group_norm": true,
60
+ "vision_backbone": "resnet18"
61
+ }
config.yaml ADDED
@@ -0,0 +1,174 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ resume: false
2
+ device: cuda
3
+ use_amp: false
4
+ seed: 100000
5
+ dataset_repo_id: hellorobotinc/kitchen_cabinet_diagonal
6
+ video_backend: pyav
7
+ training:
8
+ offline_steps: 200000
9
+ online_steps: 0
10
+ online_steps_between_rollouts: 1
11
+ online_sampling_ratio: 0.5
12
+ online_env_seed: ???
13
+ eval_freq: -1
14
+ log_freq: 250
15
+ save_checkpoint: true
16
+ save_freq: 10000
17
+ num_workers: 16
18
+ batch_size: 64
19
+ image_transforms:
20
+ enable: true
21
+ max_num_transforms: 3
22
+ random_order: false
23
+ brightness:
24
+ weight: 1
25
+ min_max:
26
+ - 0.8
27
+ - 1.2
28
+ contrast:
29
+ weight: 1
30
+ min_max:
31
+ - 0.8
32
+ - 1.2
33
+ saturation:
34
+ weight: 1
35
+ min_max:
36
+ - 0.5
37
+ - 1.5
38
+ hue:
39
+ weight: 1
40
+ min_max:
41
+ - -0.05
42
+ - 0.05
43
+ sharpness:
44
+ weight: 1
45
+ min_max:
46
+ - 0.8
47
+ - 1.2
48
+ grad_clip_norm: 10
49
+ lr: 0.0001
50
+ lr_scheduler: cosine
51
+ lr_warmup_steps: 500
52
+ adam_betas:
53
+ - 0.95
54
+ - 0.999
55
+ adam_eps: 1.0e-08
56
+ adam_weight_decay: 1.0e-06
57
+ delta_timestamps:
58
+ observation.images.gripper:
59
+ - -0.16666666666666666
60
+ - 0.0
61
+ observation.images.head:
62
+ - -0.16666666666666666
63
+ - 0.0
64
+ observation.state:
65
+ - -0.16666666666666666
66
+ - 0.0
67
+ action:
68
+ - -0.16666666666666666
69
+ - 0.0
70
+ - 0.16666666666666666
71
+ - 0.3333333333333333
72
+ - 0.5
73
+ - 0.6666666666666666
74
+ - 0.8333333333333334
75
+ - 1.0
76
+ - 1.1666666666666667
77
+ - 1.3333333333333333
78
+ - 1.5
79
+ - 1.6666666666666667
80
+ - 1.8333333333333333
81
+ - 2.0
82
+ - 2.1666666666666665
83
+ - 2.3333333333333335
84
+ drop_n_last_frames: 7
85
+ eval:
86
+ n_episodes: 50
87
+ batch_size: 50
88
+ use_async_envs: false
89
+ wandb:
90
+ enable: true
91
+ disable_artifact: true
92
+ project: diffusion-kitchen-diagonal
93
+ notes: no depth
94
+ fps: 6
95
+ env:
96
+ name: stretch_real
97
+ task: stretch_baseX
98
+ state_dim: 9
99
+ action_dim: 9
100
+ fps: ${fps}
101
+ episode_length: 400
102
+ gym:
103
+ fps: ${fps}
104
+ repo_version: main
105
+ override_dataset_stats:
106
+ observation.images.gripper:
107
+ mean:
108
+ - - - 0.485
109
+ - - - 0.456
110
+ - - - 0.406
111
+ std:
112
+ - - - 0.229
113
+ - - - 0.224
114
+ - - - 0.225
115
+ observation.images.head:
116
+ mean:
117
+ - - - 0.485
118
+ - - - 0.456
119
+ - - - 0.406
120
+ std:
121
+ - - - 0.229
122
+ - - - 0.224
123
+ - - - 0.225
124
+ policy:
125
+ name: diffusion
126
+ n_obs_steps: 2
127
+ horizon: 16
128
+ n_action_steps: 8
129
+ input_shapes:
130
+ observation.images.gripper:
131
+ - 3
132
+ - 320
133
+ - 320
134
+ observation.images.head:
135
+ - 3
136
+ - 320
137
+ - 320
138
+ observation.state:
139
+ - ${env.state_dim}
140
+ output_shapes:
141
+ action:
142
+ - ${env.action_dim}
143
+ input_normalization_modes:
144
+ observation.images.gripper: mean_std
145
+ observation.images.head: mean_std
146
+ observation.state: min_max
147
+ output_normalization_modes:
148
+ action: min_max
149
+ vision_backbone: resnet18
150
+ crop_shape:
151
+ - 320
152
+ - 320
153
+ crop_is_random: false
154
+ pretrained_backbone_weights: null
155
+ use_group_norm: true
156
+ spatial_softmax_num_keypoints: 32
157
+ down_dims:
158
+ - 512
159
+ - 1024
160
+ - 2048
161
+ kernel_size: 5
162
+ n_groups: 8
163
+ diffusion_step_embed_dim: 128
164
+ use_film_scale_modulation: true
165
+ noise_scheduler_type: DDPM
166
+ num_train_timesteps: 100
167
+ beta_schedule: squaredcos_cap_v2
168
+ beta_start: 0.0001
169
+ beta_end: 0.02
170
+ prediction_type: epsilon
171
+ clip_sample: true
172
+ clip_sample_range: 1.0
173
+ num_inference_steps: 100
174
+ do_mask_loss_for_padding: false
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e15b8e7b5ba60ff1797b0569c90dd3732c3f8c2ff18e848ea61747dc854c949f
3
+ size 1067249692