File tree Expand file tree Collapse file tree 5 files changed +73
-24
lines changed Expand file tree Collapse file tree 5 files changed +73
-24
lines changed Original file line number Diff line number Diff line change @@ -136,4 +136,5 @@ temp.gif
136
136
137
137
* .vla
138
138
* .mkv
139
- * .csv
139
+ * .csv
140
+ * .pdf
Load Diff Large diffs are not rendered by default.
Original file line number Diff line number Diff line change @@ -60,16 +60,23 @@ def __init__(
60
60
def measure_average_trajectory_size (self ):
61
61
"""Calculates the average size of trajectory files in the dataset directory."""
62
62
total_size = 0
63
- file_count = 0
64
63
for dirpath , dirnames , filenames in os .walk (self .dataset_dir ):
65
64
for f in filenames :
66
- if f .endswith (self .file_extension ):
67
- file_path = os .path .join (dirpath , f )
68
- total_size += os .path .getsize (file_path )
69
- file_count += 1
70
- if file_count == 0 :
71
- return 0
72
- return (total_size / file_count ) / (1024 * 1024 ) # Convert to MB
65
+ file_path = os .path .join (dirpath , f )
66
+ total_size += os .path .getsize (file_path )
67
+
68
+ print (f"total_size: { total_size } of directory { self .dataset_dir } " )
69
+ # trajectory number
70
+ traj_num = 0
71
+ if self .dataset_name == "nyu_door_opening_surprising_effectiveness" :
72
+ traj_num = 435
73
+ if self .dataset_name == "berkeley_cable_routing" :
74
+ traj_num = 1482
75
+ if self .dataset_name == "bridge" :
76
+ traj_num = 25460
77
+ if self .dataset_name == "berkeley_autolab_ur5" :
78
+ traj_num = 896
79
+ return (total_size / traj_num ) / (1024 * 1024 ) # Convert to MB
73
80
74
81
def clear_cache (self ):
75
82
"""Clears the cache directory."""
@@ -274,7 +281,7 @@ def __init__(
274
281
exp_dir ,
275
282
dataset_name ,
276
283
num_batches ,
277
- dataset_type = "lerobot " ,
284
+ dataset_type = "hf " ,
278
285
batch_size = batch_size ,
279
286
log_frequency = log_frequency ,
280
287
)
Original file line number Diff line number Diff line change 1
1
# ask for sudo access
2
2
sudo echo " Use sudo access for clearning cache"
3
3
4
- rm * .csv
5
-
6
4
# Define a list of batch sizes to iterate through
7
- batch_sizes=(1 2 4 6 8 )
8
- num_batches=200
5
+ batch_sizes=(1)
6
+ num_batches=20
9
7
# batch_sizes=(1 2)
10
8
11
9
# batch_sizes=(2)
Original file line number Diff line number Diff line change @@ -29,8 +29,12 @@ def _frame_to_numpy(frame):
29
29
# repeat
30
30
if self .episode_index >= len (self .dataset ):
31
31
self .episode_index = 0
32
- from_idx = self .dataset .episode_data_index ["from" ][self .episode_index ].item ()
33
- to_idx = self .dataset .episode_data_index ["to" ][self .episode_index ].item ()
32
+ try :
33
+ from_idx = self .dataset .episode_data_index ["from" ][self .episode_index ].item ()
34
+ to_idx = self .dataset .episode_data_index ["to" ][self .episode_index ].item ()
35
+ except Exception as e :
36
+ self .episode_index = 0
37
+ continue
34
38
frames = [_frame_to_numpy (self .dataset [idx ]) for idx in range (from_idx , to_idx )]
35
39
episode .extend (frames )
36
40
self .episode_index += 1
You can’t perform that action at this time.
0 commit comments