@@ -324,8 +324,10 @@ def generate_dataset(
324324 datasets , tmpdir = _determine_datasets (regions , quantifiers , chrom_size_path )
325325 # copy chrom_size_path to output_path
326326 subprocess .run (["cp" , "-f" , chrom_size_path , f"{ output_path } /chrom_sizes.txt" ], check = True )
327+ rgs = {}
327328 for region_dim , region_config in datasets .items ():
328329 regiongroup = root .create_group (region_dim )
330+ rgs [region_dim ] = regiongroup
329331 # save region coords to the ds
330332 bed = pd .read_csv (f"{ tmpdir } /{ region_dim } .regions.csv" , index_col = 0 )
331333 bed .columns = [f"{ region_dim } _chrom" , f"{ region_dim } _start" , f"{ region_dim } _end" ]
@@ -397,10 +399,9 @@ def generate_dataset(
397399 obs_dim_dtype = obs_dim_dtype ,
398400 region_dim = region_dim ,
399401 chunk_start = chunk_start ,
400- regiongroup = regiongroup ,
402+ regiongroup = rgs [ region_dim ] ,
401403 )
402404 futures [f ] = (region_dim , i )
403-
404405 for f in as_completed (futures ):
405406 region_dim , i = futures [f ]
406407 print (f"Chunk { i } of { region_dim } returned" )
@@ -415,5 +416,6 @@ def generate_dataset(
415416 "ds_sample_dim" : {region_dim : obs_dim for region_dim in datasets .keys ()},
416417 },
417418 )
418- zarr .convenience .consolidate_metadata (z )
419+ for region_dim in datasets .keys ():
420+ zarr .convenience .consolidate_metadata (f"{ output_path } /{ region_dim } " )
419421 return output_path
0 commit comments