Fix create_repo exist_ok, add metadata_update for source dataset linking
Browse files- atlas-export-remote.py +34 -24
atlas-export-remote.py
CHANGED
|
@@ -66,6 +66,7 @@ from huggingface_hub import (
|
|
| 66 |
create_repo,
|
| 67 |
get_token,
|
| 68 |
login,
|
|
|
|
| 69 |
upload_file,
|
| 70 |
upload_folder,
|
| 71 |
)
|
|
@@ -250,16 +251,14 @@ def upload_parquet_to_dataset(
|
|
| 250 |
logger.info(f"Creating/updating dataset repository: {data_repo_id}")
|
| 251 |
|
| 252 |
# Create dataset repository
|
| 253 |
-
|
| 254 |
-
|
| 255 |
-
|
| 256 |
-
|
| 257 |
-
|
| 258 |
-
|
| 259 |
-
|
| 260 |
-
|
| 261 |
-
else:
|
| 262 |
-
raise
|
| 263 |
|
| 264 |
# Upload parquet file
|
| 265 |
logger.info(
|
|
@@ -374,20 +373,15 @@ def deploy_space_with_remote_data(
|
|
| 374 |
logger.info(f"Creating Space: {repo_id}")
|
| 375 |
|
| 376 |
# Create the Space repository
|
| 377 |
-
|
| 378 |
-
|
| 379 |
-
|
| 380 |
-
|
| 381 |
-
|
| 382 |
-
|
| 383 |
-
|
| 384 |
-
|
| 385 |
-
|
| 386 |
-
except Exception as e:
|
| 387 |
-
if "already exists" in str(e):
|
| 388 |
-
logger.info(f"Space {repo_id} already exists, updating...")
|
| 389 |
-
else:
|
| 390 |
-
raise
|
| 391 |
|
| 392 |
# Create README
|
| 393 |
readme_content = create_space_readme(args, data_repo_id, parquet_size_gb)
|
|
@@ -701,6 +695,22 @@ def main():
|
|
| 701 |
args,
|
| 702 |
)
|
| 703 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 704 |
# Summary
|
| 705 |
logger.info("\n" + "=" * 50)
|
| 706 |
logger.info("Deployment Complete!")
|
|
|
|
| 66 |
create_repo,
|
| 67 |
get_token,
|
| 68 |
login,
|
| 69 |
+
metadata_update,
|
| 70 |
upload_file,
|
| 71 |
upload_folder,
|
| 72 |
)
|
|
|
|
| 251 |
logger.info(f"Creating/updating dataset repository: {data_repo_id}")
|
| 252 |
|
| 253 |
# Create dataset repository
|
| 254 |
+
create_repo(
|
| 255 |
+
data_repo_id,
|
| 256 |
+
repo_type="dataset",
|
| 257 |
+
private=private,
|
| 258 |
+
token=hf_token,
|
| 259 |
+
exist_ok=True,
|
| 260 |
+
)
|
| 261 |
+
logger.info(f"Dataset repository ready: {data_repo_id}")
|
|
|
|
|
|
|
| 262 |
|
| 263 |
# Upload parquet file
|
| 264 |
logger.info(
|
|
|
|
| 373 |
logger.info(f"Creating Space: {repo_id}")
|
| 374 |
|
| 375 |
# Create the Space repository
|
| 376 |
+
create_repo(
|
| 377 |
+
repo_id,
|
| 378 |
+
repo_type="space",
|
| 379 |
+
space_sdk="static",
|
| 380 |
+
private=private,
|
| 381 |
+
token=hf_token,
|
| 382 |
+
exist_ok=True,
|
| 383 |
+
)
|
| 384 |
+
logger.info(f"Space repository ready: {repo_id}")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 385 |
|
| 386 |
# Create README
|
| 387 |
readme_content = create_space_readme(args, data_repo_id, parquet_size_gb)
|
|
|
|
| 695 |
args,
|
| 696 |
)
|
| 697 |
|
| 698 |
+
# Link source datasets in Space metadata
|
| 699 |
+
space_repo_id = space_url.split("/spaces/")[-1]
|
| 700 |
+
dataset_ids = [data_repo_id]
|
| 701 |
+
if args.inputs:
|
| 702 |
+
for inp in args.inputs:
|
| 703 |
+
if not inp.startswith("http") and "/" in inp:
|
| 704 |
+
dataset_ids.append(inp)
|
| 705 |
+
metadata_update(
|
| 706 |
+
space_repo_id,
|
| 707 |
+
{"datasets": dataset_ids},
|
| 708 |
+
repo_type="space",
|
| 709 |
+
overwrite=True,
|
| 710 |
+
token=hf_token,
|
| 711 |
+
)
|
| 712 |
+
logger.info(f"Linked datasets in Space metadata: {dataset_ids}")
|
| 713 |
+
|
| 714 |
# Summary
|
| 715 |
logger.info("\n" + "=" * 50)
|
| 716 |
logger.info("Deployment Complete!")
|