1616 post_metadata_definition ,
1717)
1818
19- from scripts .migration .dataset_collection_json import get_dataset_collections_map
19+ from scripts .migration .dataset_collection_json import get_dataset_collections_map , get_datasets_in_collections
2020
2121DATASET_COLLECTIONS_MAP = get_dataset_collections_map ()
2222
@@ -1248,8 +1248,6 @@ def process_user_and_resources_collections(user_v1, USER_MAP, DATASET_MAP, COLLE
12481248 headers = clowder_headers_v1 , user_v1 = user_v1
12491249 )
12501250
1251- all_collection = get_clowder_v1_user_collections (headers = clowder_headers_v1 , user_v1 = user_v1 )
1252-
12531251 print (f"Got { len (user_v1_collections )} user collections in the top level" )
12541252
12551253 # filter the collections by space
@@ -1272,6 +1270,8 @@ def process_user_and_resources_collections(user_v1, USER_MAP, DATASET_MAP, COLLE
12721270 print (f"Created dataset in v2 from collection: { top_level_col ['id' ]} - { top_level_col ['name' ]} " )
12731271 COLLETIONS_MAP [top_level_col ["id" ]] = dataset_v2
12741272
1273+ datasets_in_collections_v1 = get_datasets_in_collections ()
1274+
12751275 for dataset in user_v1_datasets :
12761276 print (f"Creating dataset in v2: { dataset ['id' ]} - { dataset ['name' ]} " )
12771277 dataset_v1_id = dataset ["id" ]
@@ -1281,6 +1281,9 @@ def process_user_and_resources_collections(user_v1, USER_MAP, DATASET_MAP, COLLE
12811281 print (toml_space_ids )
12821282 print (toml_exclude_space_ids )
12831283 # Check if dataset is in the excluded dataset list
1284+ if dataset_v1_id in datasets_in_collections_v1 :
1285+ print (f"Skipping dataset { dataset_v1_id } as it is was in a collection, already migrated" )
1286+ MIGRATE_DATASET = False
12841287 if dataset_v1_id in toml_exclude_dataset_ids :
12851288 print (f"Skipping dataset { dataset_v1_id } as it is in the exclude list." )
12861289 MIGRATE_DATASET = False
@@ -1335,9 +1338,6 @@ def process_user_and_resources_collections(user_v1, USER_MAP, DATASET_MAP, COLLE
13351338 file_v2_id = download_and_upload_file_to_matching_folder (
13361339 file , dataset_v2_id , base_user_headers_v2 , matching_folder
13371340 )
1338- file_v2_id = download_and_upload_file (
1339- file , all_dataset_folders , dataset_v2_id , base_user_headers_v2
1340- )
13411341 if file_v2_id is not None :
13421342 add_file_metadata (file , file_v2_id , clowder_headers_v1 , user_headers_v2 )
13431343 # posting the collection hierarchy as metadata
0 commit comments