Skip to content

Commit c94dd6d

Browse files
committed
this should work for datasets now as well
1 parent e2be15b commit c94dd6d

File tree

1 file changed

+6
-6
lines changed

1 file changed

+6
-6
lines changed

scripts/migration/migrate.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@
1616
post_metadata_definition,
1717
)
1818

19-
from scripts.migration.dataset_collection_json import get_dataset_collections_map
19+
from scripts.migration.dataset_collection_json import get_dataset_collections_map, get_datasets_in_collections
2020

2121
DATASET_COLLECTIONS_MAP = get_dataset_collections_map()
2222

@@ -1248,8 +1248,6 @@ def process_user_and_resources_collections(user_v1, USER_MAP, DATASET_MAP, COLLE
12481248
headers=clowder_headers_v1, user_v1=user_v1
12491249
)
12501250

1251-
all_collection = get_clowder_v1_user_collections(headers=clowder_headers_v1, user_v1=user_v1)
1252-
12531251
print(f"Got {len(user_v1_collections)} user collections in the top level")
12541252

12551253
# filter the collections by space
@@ -1272,6 +1270,8 @@ def process_user_and_resources_collections(user_v1, USER_MAP, DATASET_MAP, COLLE
12721270
print(f"Created dataset in v2 from collection: {top_level_col['id']} - {top_level_col['name']}")
12731271
COLLETIONS_MAP[top_level_col["id"]] = dataset_v2
12741272

1273+
datasets_in_collections_v1 = get_datasets_in_collections()
1274+
12751275
for dataset in user_v1_datasets:
12761276
print(f"Creating dataset in v2: {dataset['id']} - {dataset['name']}")
12771277
dataset_v1_id = dataset["id"]
@@ -1281,6 +1281,9 @@ def process_user_and_resources_collections(user_v1, USER_MAP, DATASET_MAP, COLLE
12811281
print(toml_space_ids)
12821282
print(toml_exclude_space_ids)
12831283
# Check if dataset is in the excluded dataset list
1284+
if dataset_v1_id in datasets_in_collections_v1:
1285+
print(f"Skipping dataset {dataset_v1_id} as it is was in a collection, already migrated")
1286+
MIGRATE_DATASET = False
12841287
if dataset_v1_id in toml_exclude_dataset_ids:
12851288
print(f"Skipping dataset {dataset_v1_id} as it is in the exclude list.")
12861289
MIGRATE_DATASET = False
@@ -1335,9 +1338,6 @@ def process_user_and_resources_collections(user_v1, USER_MAP, DATASET_MAP, COLLE
13351338
file_v2_id = download_and_upload_file_to_matching_folder(
13361339
file, dataset_v2_id, base_user_headers_v2, matching_folder
13371340
)
1338-
file_v2_id = download_and_upload_file(
1339-
file, all_dataset_folders, dataset_v2_id, base_user_headers_v2
1340-
)
13411341
if file_v2_id is not None:
13421342
add_file_metadata(file, file_v2_id, clowder_headers_v1, user_headers_v2)
13431343
# posting the collection hierarchy as metadata

0 commit comments

Comments
 (0)